[ 479.105422] env[61970]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61970) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 479.105841] env[61970]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61970) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 479.105841] env[61970]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61970) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 479.106183] env[61970]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 479.208190] env[61970]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61970) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 479.217923] env[61970]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61970) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 479.817327] env[61970]: INFO nova.virt.driver [None req-87ee7fcd-78e9-4f2c-a4a1-fba696663a7a None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 479.887940] env[61970]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 479.888160] env[61970]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 479.888223] env[61970]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61970) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 483.008554] env[61970]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-409ca8fe-a8f7-4c09-8e3a-8bf8097536d7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.026753] env[61970]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61970) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 483.027038] env[61970]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-95989cd5-d621-4bb6-8dd9-a56f43f8a330 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.075087] env[61970]: INFO oslo_vmware.api [-] Successfully established new session; session ID is d967a. [ 483.075333] env[61970]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.187s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 483.075898] env[61970]: INFO nova.virt.vmwareapi.driver [None req-87ee7fcd-78e9-4f2c-a4a1-fba696663a7a None None] VMware vCenter version: 7.0.3 [ 483.079748] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e184ffe-9443-46b3-998b-9da161a69baf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.100095] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfbf9197-2f83-4184-81de-00c23adbcbbd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.106897] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c857405-b82a-46eb-af89-361ebb7e24b3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.113922] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2fc86c9-5580-41ef-a838-5a8e94bb7856 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.127134] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-233e0085-594e-4fa8-ae70-d081d4448293 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.133236] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a87099-80df-4d86-a24c-80429985cfff {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.162754] env[61970]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-2ed0e2cf-f810-4503-ad1b-8a8d0b57634b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.168219] env[61970]: DEBUG nova.virt.vmwareapi.driver [None req-87ee7fcd-78e9-4f2c-a4a1-fba696663a7a None None] Extension org.openstack.compute already exists. {{(pid=61970) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:226}} [ 483.170912] env[61970]: INFO nova.compute.provider_config [None req-87ee7fcd-78e9-4f2c-a4a1-fba696663a7a None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 483.675040] env[61970]: DEBUG nova.context [None req-87ee7fcd-78e9-4f2c-a4a1-fba696663a7a None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),d7823240-c2e3-4cc5-95b2-724b69bb9256(cell1) {{(pid=61970) load_cells /opt/stack/nova/nova/context.py:464}} [ 483.676649] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 483.676876] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 483.677655] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 483.678151] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] Acquiring lock "d7823240-c2e3-4cc5-95b2-724b69bb9256" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 483.678351] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] Lock "d7823240-c2e3-4cc5-95b2-724b69bb9256" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 483.679571] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] Lock "d7823240-c2e3-4cc5-95b2-724b69bb9256" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 483.699969] env[61970]: INFO dbcounter [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] Registered counter for database nova_cell0 [ 483.708074] env[61970]: INFO dbcounter [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] Registered counter for database nova_cell1 [ 483.711584] env[61970]: DEBUG oslo_db.sqlalchemy.engines [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61970) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 483.711943] env[61970]: DEBUG oslo_db.sqlalchemy.engines [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61970) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 483.716653] env[61970]: ERROR nova.db.main.api [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 483.716653] env[61970]: result = function(*args, **kwargs) [ 483.716653] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 483.716653] env[61970]: return func(*args, **kwargs) [ 483.716653] env[61970]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 483.716653] env[61970]: result = fn(*args, **kwargs) [ 483.716653] env[61970]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 483.716653] env[61970]: return f(*args, **kwargs) [ 483.716653] env[61970]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 483.716653] env[61970]: return db.service_get_minimum_version(context, binaries) [ 483.716653] env[61970]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 483.716653] env[61970]: _check_db_access() [ 483.716653] env[61970]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 483.716653] env[61970]: stacktrace = ''.join(traceback.format_stack()) [ 483.716653] env[61970]: [ 483.717950] env[61970]: ERROR nova.db.main.api [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 483.717950] env[61970]: result = function(*args, **kwargs) [ 483.717950] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 483.717950] env[61970]: return func(*args, **kwargs) [ 483.717950] env[61970]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 483.717950] env[61970]: result = fn(*args, **kwargs) [ 483.717950] env[61970]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 483.717950] env[61970]: return f(*args, **kwargs) [ 483.717950] env[61970]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 483.717950] env[61970]: return db.service_get_minimum_version(context, binaries) [ 483.717950] env[61970]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 483.717950] env[61970]: _check_db_access() [ 483.717950] env[61970]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 483.717950] env[61970]: stacktrace = ''.join(traceback.format_stack()) [ 483.717950] env[61970]: [ 483.718401] env[61970]: WARNING nova.objects.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] Failed to get minimum service version for cell d7823240-c2e3-4cc5-95b2-724b69bb9256 [ 483.718570] env[61970]: WARNING nova.objects.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 483.719119] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] Acquiring lock "singleton_lock" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 483.719307] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] Acquired lock "singleton_lock" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 483.719586] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] Releasing lock "singleton_lock" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 483.719941] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] Full set of CONF: {{(pid=61970) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 483.720103] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ******************************************************************************** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 483.720246] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] Configuration options gathered from: {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 483.720386] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 483.720583] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 483.720712] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ================================================================================ {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 483.720944] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] allow_resize_to_same_host = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.721155] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] arq_binding_timeout = 300 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.721291] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] backdoor_port = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.721419] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] backdoor_socket = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.721586] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] block_device_allocate_retries = 60 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.721754] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] block_device_allocate_retries_interval = 3 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.721924] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cert = self.pem {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.722107] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.722283] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] compute_monitors = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.722456] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] config_dir = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.722628] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] config_drive_format = iso9660 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.722763] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.722930] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] config_source = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.723117] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] console_host = devstack {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.723288] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] control_exchange = nova {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.723450] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cpu_allocation_ratio = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.723609] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] daemon = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.723779] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] debug = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.723941] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] default_access_ip_network_name = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.724129] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] default_availability_zone = nova {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.724291] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] default_ephemeral_format = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.724455] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] default_green_pool_size = 1000 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.724692] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.724861] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] default_schedule_zone = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.725042] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] disk_allocation_ratio = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.725218] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] enable_new_services = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.725403] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] enabled_apis = ['osapi_compute'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.725567] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] enabled_ssl_apis = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.725738] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] flat_injected = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.725906] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] force_config_drive = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.726155] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] force_raw_images = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.726340] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] graceful_shutdown_timeout = 5 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.726504] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] heal_instance_info_cache_interval = 60 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.726721] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] host = cpu-1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.726900] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.727079] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] initial_disk_allocation_ratio = 1.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.727245] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] initial_ram_allocation_ratio = 1.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.727459] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.727720] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] instance_build_timeout = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.727937] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] instance_delete_interval = 300 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.728127] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] instance_format = [instance: %(uuid)s] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.728343] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] instance_name_template = instance-%08x {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.728519] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] instance_usage_audit = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.728696] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] instance_usage_audit_period = month {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.728865] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.729045] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] instances_path = /opt/stack/data/nova/instances {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.729223] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] internal_service_availability_zone = internal {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.729382] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] key = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.729543] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] live_migration_retry_count = 30 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.729712] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] log_color = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.729881] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] log_config_append = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.730066] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.730240] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] log_dir = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.730402] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] log_file = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.730534] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] log_options = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.730696] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] log_rotate_interval = 1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.730870] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] log_rotate_interval_type = days {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.731049] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] log_rotation_type = none {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.731185] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.731313] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.731482] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.731647] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.731809] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.731982] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] long_rpc_timeout = 1800 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.732160] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] max_concurrent_builds = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.732352] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] max_concurrent_live_migrations = 1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.732507] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] max_concurrent_snapshots = 5 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.732663] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] max_local_block_devices = 3 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.732877] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] max_logfile_count = 30 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.733065] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] max_logfile_size_mb = 200 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.733237] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] maximum_instance_delete_attempts = 5 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.733409] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] metadata_listen = 0.0.0.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.733581] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] metadata_listen_port = 8775 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.733753] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] metadata_workers = 2 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.733915] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] migrate_max_retries = -1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.734104] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] mkisofs_cmd = genisoimage {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.734319] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] my_block_storage_ip = 10.180.1.21 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.734452] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] my_ip = 10.180.1.21 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.734615] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] network_allocate_retries = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.734793] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.734961] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] osapi_compute_listen = 0.0.0.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.735141] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] osapi_compute_listen_port = 8774 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.735314] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] osapi_compute_unique_server_name_scope = {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.735486] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] osapi_compute_workers = 2 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.735651] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] password_length = 12 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.735811] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] periodic_enable = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.735992] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] periodic_fuzzy_delay = 60 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.736196] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] pointer_model = usbtablet {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.736371] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] preallocate_images = none {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.736530] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] publish_errors = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.736660] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] pybasedir = /opt/stack/nova {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.736820] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ram_allocation_ratio = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.736981] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] rate_limit_burst = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.737168] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] rate_limit_except_level = CRITICAL {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.737332] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] rate_limit_interval = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.737492] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] reboot_timeout = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.737734] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] reclaim_instance_interval = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.737923] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] record = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.738139] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] reimage_timeout_per_gb = 60 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.738329] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] report_interval = 120 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.738500] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] rescue_timeout = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.738662] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] reserved_host_cpus = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.738824] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] reserved_host_disk_mb = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.739015] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] reserved_host_memory_mb = 512 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.739197] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] reserved_huge_pages = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.739362] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] resize_confirm_window = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.739522] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] resize_fs_using_block_device = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.739680] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] resume_guests_state_on_host_boot = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.739849] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.740016] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] rpc_response_timeout = 60 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.740182] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] run_external_periodic_tasks = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.740354] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] running_deleted_instance_action = reap {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.740514] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] running_deleted_instance_poll_interval = 1800 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.740676] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] running_deleted_instance_timeout = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.740839] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] scheduler_instance_sync_interval = 120 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.741044] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] service_down_time = 720 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.741270] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] servicegroup_driver = db {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.741439] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] shell_completion = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.741603] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] shelved_offload_time = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.741765] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] shelved_poll_interval = 3600 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.741939] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] shutdown_timeout = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.742120] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] source_is_ipv6 = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.742311] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ssl_only = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.742572] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.742744] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] sync_power_state_interval = 600 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.742909] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] sync_power_state_pool_size = 1000 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.743095] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] syslog_log_facility = LOG_USER {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.743261] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] tempdir = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.743421] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] timeout_nbd = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.743632] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] transport_url = **** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.743812] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] update_resources_interval = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.743979] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] use_cow_images = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.744162] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] use_eventlog = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.744321] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] use_journal = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.744477] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] use_json = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.744635] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] use_rootwrap_daemon = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.744796] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] use_stderr = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.744954] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] use_syslog = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.745129] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vcpu_pin_set = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.745301] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vif_plugging_is_fatal = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.745470] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vif_plugging_timeout = 300 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.745691] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] virt_mkfs = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.745971] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] volume_usage_poll_interval = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.746209] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] watch_log_file = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.746396] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] web = /usr/share/spice-html5 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 483.746590] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.746762] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.746932] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.747125] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_concurrency.disable_process_locking = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.747781] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.748027] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.748226] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.748411] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.748590] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.748763] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.748954] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.auth_strategy = keystone {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.749147] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.compute_link_prefix = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.749341] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.749524] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.dhcp_domain = novalocal {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.749698] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.enable_instance_password = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.749869] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.glance_link_prefix = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.750055] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.750242] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.750411] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.instance_list_per_project_cells = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.750576] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.list_records_by_skipping_down_cells = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.750743] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.local_metadata_per_cell = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.750918] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.max_limit = 1000 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.751108] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.metadata_cache_expiration = 15 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.751289] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.neutron_default_tenant_id = default {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.751464] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.response_validation = warn {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.751634] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.use_neutron_default_nets = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.751806] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.751972] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.752159] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.752338] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.752515] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.vendordata_dynamic_targets = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.752682] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.vendordata_jsonfile_path = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.752863] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.753075] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.backend = dogpile.cache.memcached {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.753258] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.backend_argument = **** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.753631] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.config_prefix = cache.oslo {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.753631] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.dead_timeout = 60.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.753758] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.debug_cache_backend = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.753943] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.enable_retry_client = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.754149] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.enable_socket_keepalive = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.754330] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.enabled = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.754500] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.enforce_fips_mode = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.754668] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.expiration_time = 600 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.754837] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.hashclient_retry_attempts = 2 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.755031] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.hashclient_retry_delay = 1.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.755194] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.memcache_dead_retry = 300 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.755357] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.memcache_password = **** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.755522] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.755728] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.755844] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.memcache_pool_maxsize = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.756017] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.756193] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.memcache_sasl_enabled = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.756372] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.756541] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.memcache_socket_timeout = 1.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.756704] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.memcache_username = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.756871] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.proxies = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.757049] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.redis_db = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.757219] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.redis_password = **** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.757392] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.redis_sentinel_service_name = mymaster {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.757572] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.757743] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.redis_server = localhost:6379 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.758015] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.redis_socket_timeout = 1.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.758210] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.redis_username = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.758385] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.retry_attempts = 2 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.758554] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.retry_delay = 0.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.758723] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.socket_keepalive_count = 1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.758890] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.socket_keepalive_idle = 1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.759066] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.socket_keepalive_interval = 1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.759236] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.tls_allowed_ciphers = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.759398] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.tls_cafile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.759558] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.tls_certfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.759723] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.tls_enabled = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.759885] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cache.tls_keyfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.760073] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cinder.auth_section = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.760259] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cinder.auth_type = password {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.760426] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cinder.cafile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.760607] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cinder.catalog_info = volumev3::publicURL {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.760772] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cinder.certfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.760938] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cinder.collect_timing = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.761117] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cinder.cross_az_attach = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.761285] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cinder.debug = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.761447] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cinder.endpoint_template = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.761610] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cinder.http_retries = 3 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.761774] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cinder.insecure = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.761938] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cinder.keyfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.762126] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cinder.os_region_name = RegionOne {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.762298] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cinder.split_loggers = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.762460] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cinder.timeout = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.762634] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.762798] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] compute.cpu_dedicated_set = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.762959] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] compute.cpu_shared_set = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.763143] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] compute.image_type_exclude_list = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.763310] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.763473] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] compute.max_concurrent_disk_ops = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.763636] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] compute.max_disk_devices_to_attach = -1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.763799] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.763969] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.764150] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] compute.resource_provider_association_refresh = 300 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.764315] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.764478] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] compute.shutdown_retry_interval = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.764659] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.764841] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] conductor.workers = 2 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.765032] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] console.allowed_origins = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.765203] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] console.ssl_ciphers = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.765376] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] console.ssl_minimum_version = default {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.765546] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] consoleauth.enforce_session_timeout = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.765715] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] consoleauth.token_ttl = 600 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.765897] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.cafile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.766100] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.certfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.766278] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.collect_timing = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.766441] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.connect_retries = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.766600] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.connect_retry_delay = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.766762] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.endpoint_override = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.766925] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.insecure = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.767099] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.keyfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.767268] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.max_version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.767428] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.min_version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.767587] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.region_name = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.767748] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.retriable_status_codes = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.767987] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.service_name = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.768212] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.service_type = accelerator {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.768384] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.split_loggers = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.768548] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.status_code_retries = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.768713] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.status_code_retry_delay = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.768873] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.timeout = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.769071] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.769241] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] cyborg.version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.769424] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.backend = sqlalchemy {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.769596] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.connection = **** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.769766] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.connection_debug = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.770238] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.connection_parameters = {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.770238] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.connection_recycle_time = 3600 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.770314] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.connection_trace = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.770433] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.db_inc_retry_interval = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.770601] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.db_max_retries = 20 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.770767] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.db_max_retry_interval = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.770935] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.db_retry_interval = 1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.771115] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.max_overflow = 50 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.771285] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.max_pool_size = 5 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.771448] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.max_retries = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.771620] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.771780] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.mysql_wsrep_sync_wait = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.771939] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.pool_timeout = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.772119] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.retry_interval = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.772280] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.slave_connection = **** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.772444] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.sqlite_synchronous = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.772607] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] database.use_db_reconnect = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.772789] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.backend = sqlalchemy {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.772960] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.connection = **** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.773147] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.connection_debug = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.773320] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.connection_parameters = {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.773485] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.connection_recycle_time = 3600 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.773683] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.connection_trace = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.773852] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.db_inc_retry_interval = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.774062] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.db_max_retries = 20 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.774244] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.db_max_retry_interval = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.774411] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.db_retry_interval = 1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.774577] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.max_overflow = 50 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.774741] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.max_pool_size = 5 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.774905] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.max_retries = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.775097] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.775263] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.775423] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.pool_timeout = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.775587] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.retry_interval = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.775748] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.slave_connection = **** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.775926] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] api_database.sqlite_synchronous = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.776143] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] devices.enabled_mdev_types = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.776333] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.776510] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ephemeral_storage_encryption.default_format = luks {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.776677] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ephemeral_storage_encryption.enabled = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.776847] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.777032] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.api_servers = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.777209] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.cafile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.777379] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.certfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.777546] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.collect_timing = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.777709] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.connect_retries = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.777871] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.connect_retry_delay = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.778148] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.debug = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.778346] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.default_trusted_certificate_ids = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.778519] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.enable_certificate_validation = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.778688] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.enable_rbd_download = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.778853] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.endpoint_override = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.779032] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.insecure = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.779212] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.keyfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.779376] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.max_version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.779536] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.min_version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.779704] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.num_retries = 3 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.779878] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.rbd_ceph_conf = {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.780056] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.rbd_connect_timeout = 5 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.780239] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.rbd_pool = {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.780411] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.rbd_user = {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.780575] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.region_name = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.780737] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.retriable_status_codes = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.780900] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.service_name = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.781089] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.service_type = image {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.781258] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.split_loggers = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.781425] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.status_code_retries = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.781586] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.status_code_retry_delay = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.781745] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.timeout = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.781929] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.782111] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.verify_glance_signatures = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.782279] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] glance.version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.782448] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] guestfs.debug = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.782615] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] mks.enabled = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.782974] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.783184] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] image_cache.manager_interval = 2400 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.783361] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] image_cache.precache_concurrency = 1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.783534] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] image_cache.remove_unused_base_images = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.783738] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.783915] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.784114] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] image_cache.subdirectory_name = _base {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.784297] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.api_max_retries = 60 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.784466] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.api_retry_interval = 2 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.784629] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.auth_section = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.784793] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.auth_type = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.784960] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.cafile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.785139] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.certfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.785308] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.collect_timing = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.785473] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.conductor_group = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.785634] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.connect_retries = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.785794] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.connect_retry_delay = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.785975] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.endpoint_override = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.786171] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.insecure = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.786339] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.keyfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.786502] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.max_version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.786662] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.min_version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.786831] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.peer_list = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.786992] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.region_name = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.787173] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.retriable_status_codes = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.787339] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.serial_console_state_timeout = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.787501] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.service_name = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.787674] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.service_type = baremetal {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.787835] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.shard = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.788037] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.split_loggers = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.788290] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.status_code_retries = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.788465] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.status_code_retry_delay = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.788629] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.timeout = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.788816] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.789020] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ironic.version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.789223] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.789407] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] key_manager.fixed_key = **** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.789596] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.789761] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican.barbican_api_version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.789925] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican.barbican_endpoint = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.790115] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican.barbican_endpoint_type = public {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.790284] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican.barbican_region_name = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.790447] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican.cafile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.790609] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican.certfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.790772] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican.collect_timing = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.790935] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican.insecure = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.791111] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican.keyfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.791278] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican.number_of_retries = 60 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.791440] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican.retry_delay = 1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.791605] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican.send_service_user_token = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.791767] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican.split_loggers = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.791925] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican.timeout = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.792100] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican.verify_ssl = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.792266] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican.verify_ssl_path = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.792434] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican_service_user.auth_section = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.792600] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican_service_user.auth_type = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.792762] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican_service_user.cafile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.792926] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican_service_user.certfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.793113] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican_service_user.collect_timing = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.793284] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican_service_user.insecure = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.793446] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican_service_user.keyfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.793608] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican_service_user.split_loggers = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.793767] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] barbican_service_user.timeout = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.793936] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vault.approle_role_id = **** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.794112] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vault.approle_secret_id = **** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.794289] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vault.kv_mountpoint = secret {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.794451] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vault.kv_path = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.794617] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vault.kv_version = 2 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.794778] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vault.namespace = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.794939] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vault.root_token_id = **** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.795115] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vault.ssl_ca_crt_file = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.795287] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vault.timeout = 60.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.795453] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vault.use_ssl = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.795624] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.795800] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.auth_section = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.795985] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.auth_type = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.796179] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.cafile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.796347] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.certfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.796513] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.collect_timing = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.796674] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.connect_retries = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.796833] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.connect_retry_delay = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.796994] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.endpoint_override = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.797173] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.insecure = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.797334] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.keyfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.797494] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.max_version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.797652] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.min_version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.797809] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.region_name = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.797997] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.retriable_status_codes = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.798185] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.service_name = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.798445] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.service_type = identity {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.798870] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.split_loggers = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.798870] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.status_code_retries = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.798949] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.status_code_retry_delay = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.799127] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.timeout = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.799317] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.799483] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] keystone.version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.799688] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.connection_uri = {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.799855] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.cpu_mode = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.800034] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.cpu_model_extra_flags = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.800213] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.cpu_models = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.800388] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.cpu_power_governor_high = performance {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.800558] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.cpu_power_governor_low = powersave {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.800722] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.cpu_power_management = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.800896] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.801079] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.device_detach_attempts = 8 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.801251] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.device_detach_timeout = 20 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.801420] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.disk_cachemodes = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.801584] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.disk_prefix = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.801753] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.enabled_perf_events = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.801917] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.file_backed_memory = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.802100] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.gid_maps = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.802268] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.hw_disk_discard = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.802431] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.hw_machine_type = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.802606] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.images_rbd_ceph_conf = {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.802778] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.802947] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.803135] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.images_rbd_glance_store_name = {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.803308] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.images_rbd_pool = rbd {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.803481] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.images_type = default {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.803642] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.images_volume_group = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.803806] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.inject_key = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.803973] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.inject_partition = -2 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.804153] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.inject_password = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.804322] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.iscsi_iface = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.804486] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.iser_use_multipath = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.804653] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.live_migration_bandwidth = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.804818] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.804983] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.live_migration_downtime = 500 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.805165] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.805331] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.805494] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.live_migration_inbound_addr = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.805658] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.805826] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.live_migration_permit_post_copy = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.806030] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.live_migration_scheme = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.806233] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.live_migration_timeout_action = abort {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.806405] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.live_migration_tunnelled = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.806568] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.live_migration_uri = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.806732] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.live_migration_with_native_tls = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.806892] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.max_queues = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.807071] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.807317] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.807488] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.nfs_mount_options = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.807778] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.807977] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.808174] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.num_iser_scan_tries = 5 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.808411] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.num_memory_encrypted_guests = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.808602] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.808773] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.num_pcie_ports = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.808946] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.num_volume_scan_tries = 5 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.809135] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.pmem_namespaces = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.809302] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.quobyte_client_cfg = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.809597] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.809775] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.rbd_connect_timeout = 5 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.809946] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.810132] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.810301] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.rbd_secret_uuid = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.810464] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.rbd_user = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.810630] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.810806] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.remote_filesystem_transport = ssh {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.811168] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.rescue_image_id = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.811224] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.rescue_kernel_id = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.811566] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.rescue_ramdisk_id = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.811755] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.811920] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.rx_queue_size = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.812107] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.smbfs_mount_options = {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.812391] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.812566] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.snapshot_compression = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.812729] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.snapshot_image_format = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.812950] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.813136] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.sparse_logical_volumes = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.813306] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.swtpm_enabled = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.813479] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.swtpm_group = tss {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.813649] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.swtpm_user = tss {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.813822] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.sysinfo_serial = unique {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.813983] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.tb_cache_size = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.814160] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.tx_queue_size = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.814325] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.uid_maps = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.814487] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.use_virtio_for_bridges = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.814657] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.virt_type = kvm {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.814825] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.volume_clear = zero {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.814988] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.volume_clear_size = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.815174] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.volume_use_multipath = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.815337] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.vzstorage_cache_path = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.815506] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.815675] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.vzstorage_mount_group = qemu {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.815842] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.vzstorage_mount_opts = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.816048] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.816335] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.816516] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.vzstorage_mount_user = stack {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.816686] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.816863] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.auth_section = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.817056] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.auth_type = password {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.817229] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.cafile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.817392] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.certfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.817558] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.collect_timing = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.817717] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.connect_retries = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.817877] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.connect_retry_delay = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.818099] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.default_floating_pool = public {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.818274] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.endpoint_override = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.818509] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.extension_sync_interval = 600 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.818719] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.http_retries = 3 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.818895] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.insecure = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.819086] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.keyfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.819301] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.max_version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.819481] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.819686] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.min_version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.819873] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.ovs_bridge = br-int {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.820057] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.physnets = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.820250] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.region_name = RegionOne {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.820414] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.retriable_status_codes = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.820587] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.service_metadata_proxy = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.820750] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.service_name = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.820921] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.service_type = network {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.821101] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.split_loggers = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.821265] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.status_code_retries = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.821456] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.status_code_retry_delay = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.821622] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.timeout = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.821813] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.821976] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] neutron.version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.822168] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] notifications.bdms_in_notifications = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.822350] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] notifications.default_level = INFO {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.822528] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] notifications.notification_format = unversioned {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.822693] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] notifications.notify_on_state_change = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.822872] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.823062] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] pci.alias = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.823240] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] pci.device_spec = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.823407] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] pci.report_in_placement = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.823580] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.auth_section = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.823789] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.auth_type = password {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.823965] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.824155] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.cafile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.824320] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.certfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.824484] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.collect_timing = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.824643] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.connect_retries = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.824801] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.connect_retry_delay = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.824959] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.default_domain_id = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.825133] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.default_domain_name = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.825292] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.domain_id = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.825451] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.domain_name = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.825609] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.endpoint_override = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.825772] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.insecure = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.825959] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.keyfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.826157] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.max_version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.826350] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.min_version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.826528] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.password = **** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.826690] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.project_domain_id = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.826857] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.project_domain_name = Default {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.827040] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.project_id = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.827222] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.project_name = service {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.827394] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.region_name = RegionOne {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.827556] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.retriable_status_codes = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.827718] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.service_name = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.827889] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.service_type = placement {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.828093] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.split_loggers = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.828266] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.status_code_retries = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.828430] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.status_code_retry_delay = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.828676] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.system_scope = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.828864] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.timeout = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.829066] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.trust_id = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.829249] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.user_domain_id = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.829440] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.user_domain_name = Default {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.829604] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.user_id = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.829777] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.username = nova {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.829961] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.830142] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] placement.version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.830324] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] quota.cores = 20 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.830491] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] quota.count_usage_from_placement = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.830666] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.830842] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] quota.injected_file_content_bytes = 10240 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.831020] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] quota.injected_file_path_length = 255 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.831193] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] quota.injected_files = 5 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.831363] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] quota.instances = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.831531] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] quota.key_pairs = 100 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.831698] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] quota.metadata_items = 128 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.831865] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] quota.ram = 51200 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.832039] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] quota.recheck_quota = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.832215] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] quota.server_group_members = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.832385] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] quota.server_groups = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.832583] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.832758] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.832925] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] scheduler.image_metadata_prefilter = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.833101] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.833273] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] scheduler.max_attempts = 3 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.833438] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] scheduler.max_placement_results = 1000 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.833603] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.833768] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] scheduler.query_placement_for_image_type_support = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.833929] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.834121] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] scheduler.workers = 2 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.834301] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.834474] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.834655] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.834825] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.834992] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.835176] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.835344] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.835536] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.835710] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.host_subset_size = 1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.835876] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.836083] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.836260] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.836432] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.isolated_hosts = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.836608] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.isolated_images = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.836777] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.836939] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.837120] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.837288] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.pci_in_placement = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.837452] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.837614] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.837777] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.837970] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.838181] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.838364] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.838531] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.track_instance_changes = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.838795] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.839012] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] metrics.required = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.839198] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] metrics.weight_multiplier = 1.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.839366] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.839536] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] metrics.weight_setting = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.839854] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.840042] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] serial_console.enabled = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.840233] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] serial_console.port_range = 10000:20000 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.840408] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.840579] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.840746] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] serial_console.serialproxy_port = 6083 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.840916] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] service_user.auth_section = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.841119] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] service_user.auth_type = password {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.841272] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] service_user.cafile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.841432] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] service_user.certfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.841603] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] service_user.collect_timing = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.841770] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] service_user.insecure = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.841930] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] service_user.keyfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.842117] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] service_user.send_service_user_token = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.842287] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] service_user.split_loggers = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.842464] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] service_user.timeout = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.842637] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] spice.agent_enabled = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.842802] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] spice.enabled = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.843132] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.843335] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.843509] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] spice.html5proxy_port = 6082 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.843677] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] spice.image_compression = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.843838] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] spice.jpeg_compression = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.843999] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] spice.playback_compression = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.844180] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] spice.require_secure = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.844351] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] spice.server_listen = 127.0.0.1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.844522] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.844684] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] spice.streaming_mode = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.844843] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] spice.zlib_compression = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.845014] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] upgrade_levels.baseapi = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.845199] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] upgrade_levels.compute = auto {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.845364] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] upgrade_levels.conductor = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.845525] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] upgrade_levels.scheduler = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.845690] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vendordata_dynamic_auth.auth_section = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.845852] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vendordata_dynamic_auth.auth_type = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.846103] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vendordata_dynamic_auth.cafile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.846337] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vendordata_dynamic_auth.certfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.846520] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.846687] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vendordata_dynamic_auth.insecure = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.846851] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vendordata_dynamic_auth.keyfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.847028] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.847198] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vendordata_dynamic_auth.timeout = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.847377] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.api_retry_count = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.847542] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.ca_file = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.847718] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.cache_prefix = devstack-image-cache {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.847886] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.cluster_name = testcl1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.848094] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.connection_pool_size = 10 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.848269] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.console_delay_seconds = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.848445] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.datastore_regex = ^datastore.* {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.848656] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.848934] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.host_password = **** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.849226] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.host_port = 443 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.849423] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.host_username = administrator@vsphere.local {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.849600] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.insecure = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.849767] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.integration_bridge = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.849936] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.maximum_objects = 100 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.850117] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.pbm_default_policy = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.850287] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.pbm_enabled = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.850449] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.pbm_wsdl_location = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.850622] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.850786] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.serial_port_proxy_uri = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.850948] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.serial_port_service_uri = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.851137] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.task_poll_interval = 0.5 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.851313] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.use_linked_clone = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.851486] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.vnc_keymap = en-us {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.851654] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.vnc_port = 5900 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.851820] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vmware.vnc_port_total = 10000 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.852015] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vnc.auth_schemes = ['none'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.852206] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vnc.enabled = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.852501] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.852694] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.852870] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vnc.novncproxy_port = 6080 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.853061] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vnc.server_listen = 127.0.0.1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.853243] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.853408] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vnc.vencrypt_ca_certs = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.853572] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vnc.vencrypt_client_cert = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.853734] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vnc.vencrypt_client_key = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.853928] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.854141] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.disable_deep_image_inspection = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.854313] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.854477] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.854639] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.854802] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.disable_rootwrap = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.854965] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.enable_numa_live_migration = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.855148] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.855375] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.855557] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.855722] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.libvirt_disable_apic = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.855884] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.856064] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.856235] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.856397] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.856556] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.856715] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.856873] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.857044] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.857212] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.857378] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.857563] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.857732] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] wsgi.client_socket_timeout = 900 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.857910] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] wsgi.default_pool_size = 1000 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.858111] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] wsgi.keep_alive = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.858293] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] wsgi.max_header_line = 16384 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.858460] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] wsgi.secure_proxy_ssl_header = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.858621] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] wsgi.ssl_ca_file = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.858852] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] wsgi.ssl_cert_file = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.859079] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] wsgi.ssl_key_file = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.859267] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] wsgi.tcp_keepidle = 600 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.859450] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.859622] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] zvm.ca_file = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.859785] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] zvm.cloud_connector_url = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.860090] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.860276] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] zvm.reachable_timeout = 300 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.860460] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_policy.enforce_new_defaults = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.860851] env[61970]: WARNING oslo_config.cfg [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 483.861050] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_policy.enforce_scope = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.861239] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_policy.policy_default_rule = default {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.861426] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.861605] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_policy.policy_file = policy.yaml {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.861779] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.861945] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.862125] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.862574] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.862574] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.862633] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.862773] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.862950] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] profiler.connection_string = messaging:// {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.863135] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] profiler.enabled = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.863310] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] profiler.es_doc_type = notification {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.863475] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] profiler.es_scroll_size = 10000 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.863646] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] profiler.es_scroll_time = 2m {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.863809] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] profiler.filter_error_trace = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.863980] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] profiler.hmac_keys = **** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.864169] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] profiler.sentinel_service_name = mymaster {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.864339] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] profiler.socket_timeout = 0.1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.864503] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] profiler.trace_requests = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.864666] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] profiler.trace_sqlalchemy = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.864846] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] profiler_jaeger.process_tags = {} {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.865015] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] profiler_jaeger.service_name_prefix = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.865190] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] profiler_otlp.service_name_prefix = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.865355] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] remote_debug.host = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.865516] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] remote_debug.port = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.865694] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.865857] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.866063] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.866243] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.866409] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.866572] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.866733] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.866897] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.867080] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.867260] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.867423] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.867594] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.867765] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.867961] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.868164] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.868338] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.868505] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.868683] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.868850] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.869134] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.869317] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.869489] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.869656] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.869827] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.869994] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.870177] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.870344] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.870509] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.870680] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.870849] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.ssl = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.871080] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.871214] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.871382] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.871558] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.871732] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.ssl_version = {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.871897] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.872100] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.872275] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_notifications.retry = -1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.872460] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.872639] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_messaging_notifications.transport_url = **** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.872811] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.auth_section = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.872978] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.auth_type = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.873157] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.cafile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.873318] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.certfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.873480] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.collect_timing = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.873671] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.connect_retries = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.873839] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.connect_retry_delay = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.874022] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.endpoint_id = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.874176] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.endpoint_override = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.874341] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.insecure = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.874502] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.keyfile = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.874709] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.max_version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.874884] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.min_version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.875059] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.region_name = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.875229] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.retriable_status_codes = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.875393] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.service_name = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.875553] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.service_type = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.875716] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.split_loggers = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.875877] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.status_code_retries = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.876086] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.status_code_retry_delay = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.876259] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.timeout = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.876421] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.valid_interfaces = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.876580] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_limit.version = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.876747] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_reports.file_event_handler = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.876915] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.877092] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] oslo_reports.log_dir = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.877269] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.877432] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.877592] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.877758] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.877942] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.878138] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.878572] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.878572] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vif_plug_ovs_privileged.group = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.878628] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.878783] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.879042] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.879247] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] vif_plug_ovs_privileged.user = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.879430] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] os_vif_linux_bridge.flat_interface = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.879616] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.879795] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.879971] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.880167] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.880350] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.880523] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.880689] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.880868] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.881051] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] os_vif_ovs.isolate_vif = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.881231] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.881400] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.881573] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.881743] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] os_vif_ovs.ovsdb_interface = native {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.881908] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] os_vif_ovs.per_port_bridge = False {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.882094] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] privsep_osbrick.capabilities = [21] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.882260] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] privsep_osbrick.group = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.882421] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] privsep_osbrick.helper_command = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.882587] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.882751] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.882913] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] privsep_osbrick.user = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.883101] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.883266] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] nova_sys_admin.group = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.883427] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] nova_sys_admin.helper_command = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.883592] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.883755] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.883914] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] nova_sys_admin.user = None {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 483.884058] env[61970]: DEBUG oslo_service.service [None req-1e7bac1f-f8ac-4840-ae5b-fed894048ad8 None None] ******************************************************************************** {{(pid=61970) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 483.884539] env[61970]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 484.388334] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Getting list of instances from cluster (obj){ [ 484.388334] env[61970]: value = "domain-c8" [ 484.388334] env[61970]: _type = "ClusterComputeResource" [ 484.388334] env[61970]: } {{(pid=61970) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 484.389642] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83318bc0-567e-4ccd-a186-f5bc16d3939f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 484.398935] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Got total of 0 instances {{(pid=61970) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 484.399492] env[61970]: WARNING nova.virt.vmwareapi.driver [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 484.399966] env[61970]: INFO nova.virt.node [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Generated node identity fc0f3e4e-5ef5-45e9-8335-cc68743405aa [ 484.400219] env[61970]: INFO nova.virt.node [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Wrote node identity fc0f3e4e-5ef5-45e9-8335-cc68743405aa to /opt/stack/data/n-cpu-1/compute_id [ 484.904059] env[61970]: WARNING nova.compute.manager [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Compute nodes ['fc0f3e4e-5ef5-45e9-8335-cc68743405aa'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 485.907936] env[61970]: INFO nova.compute.manager [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 486.915571] env[61970]: WARNING nova.compute.manager [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 486.915926] env[61970]: DEBUG oslo_concurrency.lockutils [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 486.916161] env[61970]: DEBUG oslo_concurrency.lockutils [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 486.916364] env[61970]: DEBUG oslo_concurrency.lockutils [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 486.916545] env[61970]: DEBUG nova.compute.resource_tracker [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61970) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 486.917487] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c2eb88-9d78-4728-b148-04a74ce2f2e6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 486.926391] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3bff8af-642f-47fa-b620-583ec83abaec {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 486.940054] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52bf9028-ddc3-41ea-afac-98b4647037e3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 486.946799] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f489018-bbab-45e1-b64c-b2fd077d7c35 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 486.975538] env[61970]: DEBUG nova.compute.resource_tracker [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181529MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61970) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 486.975698] env[61970]: DEBUG oslo_concurrency.lockutils [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 486.975861] env[61970]: DEBUG oslo_concurrency.lockutils [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 487.478633] env[61970]: WARNING nova.compute.resource_tracker [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] No compute node record for cpu-1:fc0f3e4e-5ef5-45e9-8335-cc68743405aa: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host fc0f3e4e-5ef5-45e9-8335-cc68743405aa could not be found. [ 487.982797] env[61970]: INFO nova.compute.resource_tracker [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: fc0f3e4e-5ef5-45e9-8335-cc68743405aa [ 489.491252] env[61970]: DEBUG nova.compute.resource_tracker [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 489.491588] env[61970]: DEBUG nova.compute.resource_tracker [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 489.651422] env[61970]: INFO nova.scheduler.client.report [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] [req-dea1c0fd-01c2-45ca-9f6c-c8a3f671c238] Created resource provider record via placement API for resource provider with UUID fc0f3e4e-5ef5-45e9-8335-cc68743405aa and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 489.667574] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3e07a87-bd0c-4617-8c27-3b6f4f775511 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 489.675112] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a76b5cfd-b581-4024-885d-5501ccc33e8a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 489.703789] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c720c39-e48d-411c-a1cb-6bfa2b10cfa6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 489.710487] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-822dd51c-1049-4465-9632-d9e77d6d4325 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 489.723081] env[61970]: DEBUG nova.compute.provider_tree [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 490.257154] env[61970]: DEBUG nova.scheduler.client.report [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Updated inventory for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 490.257410] env[61970]: DEBUG nova.compute.provider_tree [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Updating resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa generation from 0 to 1 during operation: update_inventory {{(pid=61970) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 490.257555] env[61970]: DEBUG nova.compute.provider_tree [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 490.312200] env[61970]: DEBUG nova.compute.provider_tree [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Updating resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa generation from 1 to 2 during operation: update_traits {{(pid=61970) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 490.817462] env[61970]: DEBUG nova.compute.resource_tracker [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 490.817826] env[61970]: DEBUG oslo_concurrency.lockutils [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.842s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 490.817950] env[61970]: DEBUG nova.service [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Creating RPC server for service compute {{(pid=61970) start /opt/stack/nova/nova/service.py:186}} [ 490.833180] env[61970]: DEBUG nova.service [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] Join ServiceGroup membership for this service compute {{(pid=61970) start /opt/stack/nova/nova/service.py:203}} [ 490.833386] env[61970]: DEBUG nova.servicegroup.drivers.db [None req-41e51c56-6a7d-4bd0-8fae-b151a90b6180 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61970) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 521.834860] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 522.338855] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Getting list of instances from cluster (obj){ [ 522.338855] env[61970]: value = "domain-c8" [ 522.338855] env[61970]: _type = "ClusterComputeResource" [ 522.338855] env[61970]: } {{(pid=61970) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 522.340489] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6a52009-5810-4d9f-9bb0-e5992df8ad38 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.350462] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Got total of 0 instances {{(pid=61970) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 522.350858] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 522.351234] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Getting list of instances from cluster (obj){ [ 522.351234] env[61970]: value = "domain-c8" [ 522.351234] env[61970]: _type = "ClusterComputeResource" [ 522.351234] env[61970]: } {{(pid=61970) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 522.352376] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e7dcff-7df9-4fa3-83a7-7ebb2ae1efcb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.359650] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Got total of 0 instances {{(pid=61970) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 529.218497] env[61970]: DEBUG oslo_concurrency.lockutils [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Acquiring lock "3ff41f60-357c-4b56-b106-0e69207b78a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.219241] env[61970]: DEBUG oslo_concurrency.lockutils [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Lock "3ff41f60-357c-4b56-b106-0e69207b78a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.723737] env[61970]: DEBUG nova.compute.manager [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 530.272985] env[61970]: DEBUG oslo_concurrency.lockutils [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.272985] env[61970]: DEBUG oslo_concurrency.lockutils [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.274457] env[61970]: INFO nova.compute.claims [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 530.785992] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "10e1500f-70f1-40b6-aad6-5f9d75e1554b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.786381] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "10e1500f-70f1-40b6-aad6-5f9d75e1554b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.290857] env[61970]: DEBUG nova.compute.manager [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 531.366252] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16320d49-d49a-4872-95c3-43c1497c62bc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.377761] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-677df53b-3a02-445f-b62f-1fdc846fdb68 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.414249] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-043c9452-c66f-4fe9-be9e-1245d37839f4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.422342] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-059873f3-d67f-4c7e-a386-6583f387dcc8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.434124] env[61970]: DEBUG nova.compute.provider_tree [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 531.819080] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.938710] env[61970]: DEBUG nova.scheduler.client.report [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 531.966639] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Acquiring lock "0b34c674-6ef0-4c30-83e9-b4a5623ea044" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.966877] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Lock "0b34c674-6ef0-4c30-83e9-b4a5623ea044" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.444086] env[61970]: DEBUG oslo_concurrency.lockutils [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.171s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.444086] env[61970]: DEBUG nova.compute.manager [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 532.450196] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.628s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.453991] env[61970]: INFO nova.compute.claims [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 532.472751] env[61970]: DEBUG nova.compute.manager [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 532.693198] env[61970]: DEBUG oslo_concurrency.lockutils [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Acquiring lock "4d9ee781-bca8-4cad-9fc5-87cb785d478e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.693429] env[61970]: DEBUG oslo_concurrency.lockutils [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Lock "4d9ee781-bca8-4cad-9fc5-87cb785d478e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.960748] env[61970]: DEBUG nova.compute.utils [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 532.964014] env[61970]: DEBUG nova.compute.manager [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 532.964291] env[61970]: DEBUG nova.network.neutron [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 533.004427] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.197819] env[61970]: DEBUG nova.compute.manager [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 533.233706] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Acquiring lock "c4d3b145-84be-4d19-afc7-4c29fa193760" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.233706] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Lock "c4d3b145-84be-4d19-afc7-4c29fa193760" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.464850] env[61970]: DEBUG nova.compute.manager [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 533.588197] env[61970]: DEBUG nova.policy [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0c17548e57454c969873252035487f69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5601cf4cffea4aa594bc541a39c4c41b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 533.635278] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece1dc74-fcbd-4011-a860-7f6d7ef222be {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.648847] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5419e2c-7380-4a0f-ab1f-e873fb1f3a79 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.685039] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91f0b83b-d2eb-4625-88d5-9510d61f05e3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.696121] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-028e1be5-d986-4665-97c0-447e538e66d1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.715524] env[61970]: DEBUG nova.compute.provider_tree [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 533.735332] env[61970]: DEBUG oslo_concurrency.lockutils [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.735991] env[61970]: DEBUG nova.compute.manager [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 533.789932] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Acquiring lock "bce7d001-2885-409c-b455-4c8a3ce8eb9e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.790513] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Lock "bce7d001-2885-409c-b455-4c8a3ce8eb9e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.219032] env[61970]: DEBUG nova.scheduler.client.report [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 534.268318] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.295176] env[61970]: DEBUG nova.compute.manager [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 534.480775] env[61970]: DEBUG nova.compute.manager [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 534.515161] env[61970]: DEBUG nova.virt.hardware [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 534.515370] env[61970]: DEBUG nova.virt.hardware [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 534.515500] env[61970]: DEBUG nova.virt.hardware [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 534.515681] env[61970]: DEBUG nova.virt.hardware [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 534.515824] env[61970]: DEBUG nova.virt.hardware [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 534.515974] env[61970]: DEBUG nova.virt.hardware [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 534.517285] env[61970]: DEBUG nova.virt.hardware [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 534.517285] env[61970]: DEBUG nova.virt.hardware [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 534.517285] env[61970]: DEBUG nova.virt.hardware [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 534.517473] env[61970]: DEBUG nova.virt.hardware [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 534.517627] env[61970]: DEBUG nova.virt.hardware [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 534.518878] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec97354a-add9-4402-bf04-b556b76847b0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.530929] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d7977d0-d4b5-44fe-a571-f9c2f72ce165 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.535822] env[61970]: DEBUG nova.network.neutron [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Successfully created port: 9644b93e-8dd4-499b-bc7b-cc741973acc2 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 534.552312] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc220ad8-33e0-46d5-80cf-d5bf0b94e36c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.726591] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.280s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.727454] env[61970]: DEBUG nova.compute.manager [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 534.732905] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.728s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.733441] env[61970]: INFO nova.compute.claims [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 534.831332] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.236132] env[61970]: DEBUG nova.compute.utils [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 535.242144] env[61970]: DEBUG nova.compute.manager [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 535.242144] env[61970]: DEBUG nova.network.neutron [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 535.352726] env[61970]: DEBUG nova.policy [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '13aacb75013d408980ba5189c6071447', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c5b87b23cfe747379a4f42e6879f0f25', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 535.740691] env[61970]: DEBUG nova.compute.manager [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 535.888720] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25860d2c-19a1-470c-800d-612269350989 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.898717] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f57458-9755-4f9d-9451-b6d339097de8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.942183] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57c30673-7dbb-408f-be9c-fa319d1efab9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.951871] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b4bf4d9-d552-4cbe-89ed-e40beb1896c2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.968150] env[61970]: DEBUG nova.compute.provider_tree [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 536.131404] env[61970]: DEBUG nova.network.neutron [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Successfully created port: 6f11db79-b4bf-466c-902c-953c11ed548b {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 536.472222] env[61970]: DEBUG nova.scheduler.client.report [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 536.750088] env[61970]: DEBUG nova.compute.manager [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 536.784656] env[61970]: DEBUG nova.virt.hardware [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 536.784656] env[61970]: DEBUG nova.virt.hardware [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 536.784656] env[61970]: DEBUG nova.virt.hardware [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 536.784878] env[61970]: DEBUG nova.virt.hardware [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 536.784952] env[61970]: DEBUG nova.virt.hardware [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 536.786427] env[61970]: DEBUG nova.virt.hardware [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 536.786639] env[61970]: DEBUG nova.virt.hardware [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 536.787089] env[61970]: DEBUG nova.virt.hardware [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 536.787346] env[61970]: DEBUG nova.virt.hardware [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 536.787602] env[61970]: DEBUG nova.virt.hardware [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 536.787711] env[61970]: DEBUG nova.virt.hardware [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 536.788627] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aebdaaaa-f913-4aa0-9ac2-7759ddffc863 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.799874] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-670f3ad2-0bfb-4d36-bdac-092cd210887b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.984525] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.253s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 536.988166] env[61970]: DEBUG nova.compute.manager [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 536.989109] env[61970]: DEBUG oslo_concurrency.lockutils [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.254s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.992643] env[61970]: INFO nova.compute.claims [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 537.499025] env[61970]: DEBUG nova.compute.utils [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 537.501131] env[61970]: DEBUG nova.compute.manager [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 537.501425] env[61970]: DEBUG nova.network.neutron [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 537.666818] env[61970]: DEBUG nova.policy [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f314e9220e2646e99effa0833275764b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c90477b70a08498f9100f8de6f2e7191', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 537.801220] env[61970]: ERROR nova.compute.manager [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9644b93e-8dd4-499b-bc7b-cc741973acc2, please check neutron logs for more information. [ 537.801220] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 537.801220] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.801220] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 537.801220] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.801220] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 537.801220] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.801220] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 537.801220] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.801220] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 537.801220] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.801220] env[61970]: ERROR nova.compute.manager raise self.value [ 537.801220] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.801220] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 537.801220] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.801220] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 537.801846] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.801846] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 537.801846] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9644b93e-8dd4-499b-bc7b-cc741973acc2, please check neutron logs for more information. [ 537.801846] env[61970]: ERROR nova.compute.manager [ 537.801846] env[61970]: Traceback (most recent call last): [ 537.801846] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 537.801846] env[61970]: listener.cb(fileno) [ 537.801846] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.801846] env[61970]: result = function(*args, **kwargs) [ 537.801846] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 537.801846] env[61970]: return func(*args, **kwargs) [ 537.801846] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.801846] env[61970]: raise e [ 537.801846] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.801846] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 537.801846] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.801846] env[61970]: created_port_ids = self._update_ports_for_instance( [ 537.801846] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.801846] env[61970]: with excutils.save_and_reraise_exception(): [ 537.801846] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.801846] env[61970]: self.force_reraise() [ 537.801846] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.801846] env[61970]: raise self.value [ 537.801846] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.801846] env[61970]: updated_port = self._update_port( [ 537.801846] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.801846] env[61970]: _ensure_no_port_binding_failure(port) [ 537.801846] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.801846] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 537.802954] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 9644b93e-8dd4-499b-bc7b-cc741973acc2, please check neutron logs for more information. [ 537.802954] env[61970]: Removing descriptor: 15 [ 537.807340] env[61970]: ERROR nova.compute.manager [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9644b93e-8dd4-499b-bc7b-cc741973acc2, please check neutron logs for more information. [ 537.807340] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Traceback (most recent call last): [ 537.807340] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 537.807340] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] yield resources [ 537.807340] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 537.807340] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] self.driver.spawn(context, instance, image_meta, [ 537.807340] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 537.807340] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 537.807340] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 537.807340] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] vm_ref = self.build_virtual_machine(instance, [ 537.807340] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 537.807939] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] vif_infos = vmwarevif.get_vif_info(self._session, [ 537.807939] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 537.807939] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] for vif in network_info: [ 537.807939] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 537.807939] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] return self._sync_wrapper(fn, *args, **kwargs) [ 537.807939] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 537.807939] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] self.wait() [ 537.807939] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 537.807939] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] self[:] = self._gt.wait() [ 537.807939] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 537.807939] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] return self._exit_event.wait() [ 537.807939] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 537.807939] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] result = hub.switch() [ 537.808397] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 537.808397] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] return self.greenlet.switch() [ 537.808397] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.808397] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] result = function(*args, **kwargs) [ 537.808397] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 537.808397] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] return func(*args, **kwargs) [ 537.808397] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.808397] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] raise e [ 537.808397] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.808397] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] nwinfo = self.network_api.allocate_for_instance( [ 537.808397] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.808397] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] created_port_ids = self._update_ports_for_instance( [ 537.808397] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.808733] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] with excutils.save_and_reraise_exception(): [ 537.808733] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.808733] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] self.force_reraise() [ 537.808733] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.808733] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] raise self.value [ 537.808733] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.808733] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] updated_port = self._update_port( [ 537.808733] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.808733] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] _ensure_no_port_binding_failure(port) [ 537.808733] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.808733] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] raise exception.PortBindingFailed(port_id=port['id']) [ 537.808733] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] nova.exception.PortBindingFailed: Binding failed for port 9644b93e-8dd4-499b-bc7b-cc741973acc2, please check neutron logs for more information. [ 537.808733] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] [ 537.809097] env[61970]: INFO nova.compute.manager [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Terminating instance [ 537.812032] env[61970]: DEBUG oslo_concurrency.lockutils [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Acquiring lock "refresh_cache-3ff41f60-357c-4b56-b106-0e69207b78a6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.813068] env[61970]: DEBUG oslo_concurrency.lockutils [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Acquired lock "refresh_cache-3ff41f60-357c-4b56-b106-0e69207b78a6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.813068] env[61970]: DEBUG nova.network.neutron [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 538.001772] env[61970]: DEBUG nova.compute.manager [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 538.026109] env[61970]: ERROR nova.compute.manager [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6f11db79-b4bf-466c-902c-953c11ed548b, please check neutron logs for more information. [ 538.026109] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 538.026109] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.026109] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 538.026109] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.026109] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 538.026109] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.026109] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 538.026109] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.026109] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 538.026109] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.026109] env[61970]: ERROR nova.compute.manager raise self.value [ 538.026109] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.026109] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 538.026109] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.026109] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 538.026731] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.026731] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 538.026731] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6f11db79-b4bf-466c-902c-953c11ed548b, please check neutron logs for more information. [ 538.026731] env[61970]: ERROR nova.compute.manager [ 538.026731] env[61970]: Traceback (most recent call last): [ 538.026731] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 538.026731] env[61970]: listener.cb(fileno) [ 538.026731] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.026731] env[61970]: result = function(*args, **kwargs) [ 538.026731] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 538.026731] env[61970]: return func(*args, **kwargs) [ 538.026731] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.026731] env[61970]: raise e [ 538.026731] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.026731] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 538.026731] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.026731] env[61970]: created_port_ids = self._update_ports_for_instance( [ 538.026731] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.026731] env[61970]: with excutils.save_and_reraise_exception(): [ 538.026731] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.026731] env[61970]: self.force_reraise() [ 538.026731] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.026731] env[61970]: raise self.value [ 538.026731] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.026731] env[61970]: updated_port = self._update_port( [ 538.026731] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.026731] env[61970]: _ensure_no_port_binding_failure(port) [ 538.026731] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.026731] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 538.027837] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 6f11db79-b4bf-466c-902c-953c11ed548b, please check neutron logs for more information. [ 538.027837] env[61970]: Removing descriptor: 16 [ 538.027837] env[61970]: ERROR nova.compute.manager [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6f11db79-b4bf-466c-902c-953c11ed548b, please check neutron logs for more information. [ 538.027837] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Traceback (most recent call last): [ 538.027837] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 538.027837] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] yield resources [ 538.027837] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 538.027837] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] self.driver.spawn(context, instance, image_meta, [ 538.027837] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 538.027837] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 538.027837] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 538.027837] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] vm_ref = self.build_virtual_machine(instance, [ 538.028282] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 538.028282] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] vif_infos = vmwarevif.get_vif_info(self._session, [ 538.028282] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 538.028282] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] for vif in network_info: [ 538.028282] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 538.028282] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] return self._sync_wrapper(fn, *args, **kwargs) [ 538.028282] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 538.028282] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] self.wait() [ 538.028282] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 538.028282] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] self[:] = self._gt.wait() [ 538.028282] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 538.028282] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] return self._exit_event.wait() [ 538.028282] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 538.028601] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] result = hub.switch() [ 538.028601] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 538.028601] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] return self.greenlet.switch() [ 538.028601] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.028601] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] result = function(*args, **kwargs) [ 538.028601] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 538.028601] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] return func(*args, **kwargs) [ 538.028601] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.028601] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] raise e [ 538.028601] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.028601] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] nwinfo = self.network_api.allocate_for_instance( [ 538.028601] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.028601] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] created_port_ids = self._update_ports_for_instance( [ 538.029010] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.029010] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] with excutils.save_and_reraise_exception(): [ 538.029010] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.029010] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] self.force_reraise() [ 538.029010] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.029010] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] raise self.value [ 538.029010] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.029010] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] updated_port = self._update_port( [ 538.029010] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.029010] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] _ensure_no_port_binding_failure(port) [ 538.029010] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.029010] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] raise exception.PortBindingFailed(port_id=port['id']) [ 538.029401] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] nova.exception.PortBindingFailed: Binding failed for port 6f11db79-b4bf-466c-902c-953c11ed548b, please check neutron logs for more information. [ 538.029401] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] [ 538.029401] env[61970]: INFO nova.compute.manager [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Terminating instance [ 538.034463] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "refresh_cache-10e1500f-70f1-40b6-aad6-5f9d75e1554b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.035261] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquired lock "refresh_cache-10e1500f-70f1-40b6-aad6-5f9d75e1554b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.035261] env[61970]: DEBUG nova.network.neutron [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 538.167227] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f16802c-8e20-4a12-859c-5c788e9804eb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.177322] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-803f7b07-70b3-42ee-98a1-28d9d153540c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.212744] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b6f3f5-ed25-4ff9-bdd2-0eaae274dcb7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.226057] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e65e3fac-c53f-41e8-839c-2db70c004a25 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.247428] env[61970]: DEBUG nova.compute.provider_tree [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 538.412767] env[61970]: DEBUG nova.network.neutron [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.585759] env[61970]: DEBUG nova.network.neutron [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.652637] env[61970]: DEBUG nova.network.neutron [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.683780] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Acquiring lock "36c69772-6e47-4c00-a582-b46d233e13a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.683901] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Lock "36c69772-6e47-4c00-a582-b46d233e13a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.753315] env[61970]: DEBUG nova.scheduler.client.report [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 538.887143] env[61970]: DEBUG oslo_concurrency.lockutils [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Acquiring lock "e3744562-c9ab-4902-880c-6f0153b8a183" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.887426] env[61970]: DEBUG oslo_concurrency.lockutils [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Lock "e3744562-c9ab-4902-880c-6f0153b8a183" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.018689] env[61970]: DEBUG nova.compute.manager [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 539.052675] env[61970]: DEBUG nova.virt.hardware [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 539.053132] env[61970]: DEBUG nova.virt.hardware [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 539.053132] env[61970]: DEBUG nova.virt.hardware [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 539.053284] env[61970]: DEBUG nova.virt.hardware [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 539.053383] env[61970]: DEBUG nova.virt.hardware [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 539.053608] env[61970]: DEBUG nova.virt.hardware [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 539.053747] env[61970]: DEBUG nova.virt.hardware [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 539.054065] env[61970]: DEBUG nova.virt.hardware [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 539.055128] env[61970]: DEBUG nova.virt.hardware [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 539.055404] env[61970]: DEBUG nova.virt.hardware [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 539.055502] env[61970]: DEBUG nova.virt.hardware [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 539.057270] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6582fbe-2af3-4e9a-9752-ad2eb9bb5bec {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.065909] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da671e0a-2a14-4495-89ed-b5f717aaa18b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.089207] env[61970]: DEBUG oslo_concurrency.lockutils [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Releasing lock "refresh_cache-3ff41f60-357c-4b56-b106-0e69207b78a6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.089635] env[61970]: DEBUG nova.compute.manager [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 539.089831] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 539.090120] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9e160287-aa8e-430f-a9cf-6227ea1dc36b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.094669] env[61970]: DEBUG nova.network.neutron [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Successfully created port: 42772f00-251f-48ec-9d2a-b484bbf44126 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 539.100562] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac36d5d-84e2-4194-8432-52a631939f08 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.122666] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3ff41f60-357c-4b56-b106-0e69207b78a6 could not be found. [ 539.123078] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 539.123768] env[61970]: INFO nova.compute.manager [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 539.124211] env[61970]: DEBUG oslo.service.loopingcall [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 539.124611] env[61970]: DEBUG nova.compute.manager [-] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 539.124901] env[61970]: DEBUG nova.network.neutron [-] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 539.160370] env[61970]: DEBUG nova.network.neutron [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.178462] env[61970]: DEBUG nova.network.neutron [-] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.188726] env[61970]: DEBUG nova.compute.manager [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 539.262020] env[61970]: DEBUG oslo_concurrency.lockutils [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.273s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.262373] env[61970]: DEBUG nova.compute.manager [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 539.267977] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.999s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.269029] env[61970]: INFO nova.compute.claims [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 539.271606] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 539.273246] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 539.273246] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Starting heal instance info cache {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 539.273246] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Rebuilding the list of instances to heal {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 539.391044] env[61970]: DEBUG nova.compute.manager [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 539.480678] env[61970]: DEBUG nova.compute.manager [req-d36c1bcd-8bd7-4c9e-9aac-a173288ae36e req-a41885b8-752a-4a49-9e86-a93687b6db63 service nova] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Received event network-changed-9644b93e-8dd4-499b-bc7b-cc741973acc2 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.480873] env[61970]: DEBUG nova.compute.manager [req-d36c1bcd-8bd7-4c9e-9aac-a173288ae36e req-a41885b8-752a-4a49-9e86-a93687b6db63 service nova] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Refreshing instance network info cache due to event network-changed-9644b93e-8dd4-499b-bc7b-cc741973acc2. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 539.481100] env[61970]: DEBUG oslo_concurrency.lockutils [req-d36c1bcd-8bd7-4c9e-9aac-a173288ae36e req-a41885b8-752a-4a49-9e86-a93687b6db63 service nova] Acquiring lock "refresh_cache-3ff41f60-357c-4b56-b106-0e69207b78a6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.481241] env[61970]: DEBUG oslo_concurrency.lockutils [req-d36c1bcd-8bd7-4c9e-9aac-a173288ae36e req-a41885b8-752a-4a49-9e86-a93687b6db63 service nova] Acquired lock "refresh_cache-3ff41f60-357c-4b56-b106-0e69207b78a6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.481461] env[61970]: DEBUG nova.network.neutron [req-d36c1bcd-8bd7-4c9e-9aac-a173288ae36e req-a41885b8-752a-4a49-9e86-a93687b6db63 service nova] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Refreshing network info cache for port 9644b93e-8dd4-499b-bc7b-cc741973acc2 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 539.664944] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Releasing lock "refresh_cache-10e1500f-70f1-40b6-aad6-5f9d75e1554b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.665546] env[61970]: DEBUG nova.compute.manager [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 539.667967] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 539.667967] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6ce84185-c793-475d-bab0-3a9cb96d1080 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.683145] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a8ae65d-73cf-4b9f-a4bd-5bbaaa4d3393 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.693383] env[61970]: DEBUG nova.network.neutron [-] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.715214] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 10e1500f-70f1-40b6-aad6-5f9d75e1554b could not be found. [ 539.715441] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 539.715649] env[61970]: INFO nova.compute.manager [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Took 0.05 seconds to destroy the instance on the hypervisor. [ 539.715927] env[61970]: DEBUG oslo.service.loopingcall [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 539.716768] env[61970]: DEBUG nova.compute.manager [-] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 539.716864] env[61970]: DEBUG nova.network.neutron [-] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 539.721886] env[61970]: DEBUG nova.compute.manager [req-295645a7-e66e-454a-b520-4a0585177038 req-a8555883-4f82-4ec7-b5be-044324f42418 service nova] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Received event network-changed-6f11db79-b4bf-466c-902c-953c11ed548b {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.721966] env[61970]: DEBUG nova.compute.manager [req-295645a7-e66e-454a-b520-4a0585177038 req-a8555883-4f82-4ec7-b5be-044324f42418 service nova] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Refreshing instance network info cache due to event network-changed-6f11db79-b4bf-466c-902c-953c11ed548b. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 539.723146] env[61970]: DEBUG oslo_concurrency.lockutils [req-295645a7-e66e-454a-b520-4a0585177038 req-a8555883-4f82-4ec7-b5be-044324f42418 service nova] Acquiring lock "refresh_cache-10e1500f-70f1-40b6-aad6-5f9d75e1554b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.723146] env[61970]: DEBUG oslo_concurrency.lockutils [req-295645a7-e66e-454a-b520-4a0585177038 req-a8555883-4f82-4ec7-b5be-044324f42418 service nova] Acquired lock "refresh_cache-10e1500f-70f1-40b6-aad6-5f9d75e1554b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.723146] env[61970]: DEBUG nova.network.neutron [req-295645a7-e66e-454a-b520-4a0585177038 req-a8555883-4f82-4ec7-b5be-044324f42418 service nova] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Refreshing network info cache for port 6f11db79-b4bf-466c-902c-953c11ed548b {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 539.727065] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.751955] env[61970]: DEBUG nova.network.neutron [-] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.775138] env[61970]: DEBUG nova.compute.utils [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 539.775138] env[61970]: DEBUG nova.compute.manager [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 539.775138] env[61970]: DEBUG nova.network.neutron [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 539.781713] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 539.782196] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 539.782455] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 539.783348] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 539.783348] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Didn't find any instances for network info cache update. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 539.783968] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 539.784946] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 539.785282] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 539.785597] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 539.785973] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 539.786296] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 539.786913] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61970) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 539.786913] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 539.857148] env[61970]: DEBUG nova.policy [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '458c41355dfa41cf88eefafd5250c62e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9f01f3d7f46947bf930b20c7a088e599', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 539.917933] env[61970]: DEBUG oslo_concurrency.lockutils [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.037524] env[61970]: DEBUG nova.network.neutron [req-d36c1bcd-8bd7-4c9e-9aac-a173288ae36e req-a41885b8-752a-4a49-9e86-a93687b6db63 service nova] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.159943] env[61970]: DEBUG nova.network.neutron [req-d36c1bcd-8bd7-4c9e-9aac-a173288ae36e req-a41885b8-752a-4a49-9e86-a93687b6db63 service nova] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.200125] env[61970]: INFO nova.compute.manager [-] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Took 1.07 seconds to deallocate network for instance. [ 540.202888] env[61970]: DEBUG nova.compute.claims [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 540.203024] env[61970]: DEBUG oslo_concurrency.lockutils [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.254984] env[61970]: DEBUG nova.network.neutron [req-295645a7-e66e-454a-b520-4a0585177038 req-a8555883-4f82-4ec7-b5be-044324f42418 service nova] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.256814] env[61970]: DEBUG nova.network.neutron [-] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.289808] env[61970]: DEBUG nova.compute.manager [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 540.293223] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.451174] env[61970]: DEBUG nova.network.neutron [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Successfully created port: e544516b-8f51-45de-9462-e55eae71c56b {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 540.458991] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2072254e-9060-4d64-b813-692a2c69e1b7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.466851] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282d7ce5-f470-4b4f-b464-529ce91ecb17 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.503036] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc27dde7-e26f-49fc-a3c5-31ae4b69449d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.512035] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0e16eb7-24a1-454a-ae26-e909f76d83f1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.523828] env[61970]: DEBUG nova.compute.provider_tree [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 540.544959] env[61970]: DEBUG nova.network.neutron [req-295645a7-e66e-454a-b520-4a0585177038 req-a8555883-4f82-4ec7-b5be-044324f42418 service nova] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.663075] env[61970]: DEBUG oslo_concurrency.lockutils [req-d36c1bcd-8bd7-4c9e-9aac-a173288ae36e req-a41885b8-752a-4a49-9e86-a93687b6db63 service nova] Releasing lock "refresh_cache-3ff41f60-357c-4b56-b106-0e69207b78a6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.760288] env[61970]: INFO nova.compute.manager [-] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Took 1.04 seconds to deallocate network for instance. [ 540.763587] env[61970]: DEBUG nova.compute.claims [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 540.764017] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.027629] env[61970]: DEBUG nova.scheduler.client.report [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 541.048552] env[61970]: DEBUG oslo_concurrency.lockutils [req-295645a7-e66e-454a-b520-4a0585177038 req-a8555883-4f82-4ec7-b5be-044324f42418 service nova] Releasing lock "refresh_cache-10e1500f-70f1-40b6-aad6-5f9d75e1554b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.304117] env[61970]: DEBUG nova.compute.manager [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 541.343338] env[61970]: DEBUG nova.virt.hardware [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 541.343338] env[61970]: DEBUG nova.virt.hardware [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 541.343517] env[61970]: DEBUG nova.virt.hardware [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 541.343658] env[61970]: DEBUG nova.virt.hardware [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 541.343811] env[61970]: DEBUG nova.virt.hardware [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 541.344535] env[61970]: DEBUG nova.virt.hardware [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 541.344535] env[61970]: DEBUG nova.virt.hardware [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 541.344535] env[61970]: DEBUG nova.virt.hardware [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 541.344535] env[61970]: DEBUG nova.virt.hardware [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 541.344697] env[61970]: DEBUG nova.virt.hardware [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 541.344804] env[61970]: DEBUG nova.virt.hardware [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 541.347773] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a2c003b-1aff-4c31-b1f6-69780b53d3cf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.358018] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf40701d-3718-4307-83ef-82895d45bbbf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.533223] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.267s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.533745] env[61970]: DEBUG nova.compute.manager [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 541.540256] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.708s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.542116] env[61970]: INFO nova.compute.claims [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 541.680111] env[61970]: ERROR nova.compute.manager [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 42772f00-251f-48ec-9d2a-b484bbf44126, please check neutron logs for more information. [ 541.680111] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 541.680111] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.680111] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 541.680111] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.680111] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 541.680111] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.680111] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 541.680111] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.680111] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 541.680111] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.680111] env[61970]: ERROR nova.compute.manager raise self.value [ 541.680111] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.680111] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 541.680111] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.680111] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 541.680630] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.680630] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 541.680630] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 42772f00-251f-48ec-9d2a-b484bbf44126, please check neutron logs for more information. [ 541.680630] env[61970]: ERROR nova.compute.manager [ 541.680630] env[61970]: Traceback (most recent call last): [ 541.680630] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 541.680630] env[61970]: listener.cb(fileno) [ 541.680630] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.680630] env[61970]: result = function(*args, **kwargs) [ 541.680630] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 541.680630] env[61970]: return func(*args, **kwargs) [ 541.680630] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 541.680630] env[61970]: raise e [ 541.680630] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.680630] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 541.680630] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.680630] env[61970]: created_port_ids = self._update_ports_for_instance( [ 541.680630] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.680630] env[61970]: with excutils.save_and_reraise_exception(): [ 541.680630] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.680630] env[61970]: self.force_reraise() [ 541.680630] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.680630] env[61970]: raise self.value [ 541.680630] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.680630] env[61970]: updated_port = self._update_port( [ 541.680630] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.680630] env[61970]: _ensure_no_port_binding_failure(port) [ 541.680630] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.680630] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 541.681570] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 42772f00-251f-48ec-9d2a-b484bbf44126, please check neutron logs for more information. [ 541.681570] env[61970]: Removing descriptor: 17 [ 541.681570] env[61970]: ERROR nova.compute.manager [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 42772f00-251f-48ec-9d2a-b484bbf44126, please check neutron logs for more information. [ 541.681570] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Traceback (most recent call last): [ 541.681570] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 541.681570] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] yield resources [ 541.681570] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 541.681570] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] self.driver.spawn(context, instance, image_meta, [ 541.681570] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 541.681570] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] self._vmops.spawn(context, instance, image_meta, injected_files, [ 541.681570] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 541.681570] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] vm_ref = self.build_virtual_machine(instance, [ 541.682032] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 541.682032] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] vif_infos = vmwarevif.get_vif_info(self._session, [ 541.682032] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 541.682032] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] for vif in network_info: [ 541.682032] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 541.682032] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] return self._sync_wrapper(fn, *args, **kwargs) [ 541.682032] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 541.682032] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] self.wait() [ 541.682032] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 541.682032] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] self[:] = self._gt.wait() [ 541.682032] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 541.682032] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] return self._exit_event.wait() [ 541.682032] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 541.682365] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] result = hub.switch() [ 541.682365] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 541.682365] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] return self.greenlet.switch() [ 541.682365] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.682365] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] result = function(*args, **kwargs) [ 541.682365] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 541.682365] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] return func(*args, **kwargs) [ 541.682365] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 541.682365] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] raise e [ 541.682365] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.682365] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] nwinfo = self.network_api.allocate_for_instance( [ 541.682365] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.682365] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] created_port_ids = self._update_ports_for_instance( [ 541.682705] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.682705] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] with excutils.save_and_reraise_exception(): [ 541.682705] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.682705] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] self.force_reraise() [ 541.682705] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.682705] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] raise self.value [ 541.682705] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.682705] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] updated_port = self._update_port( [ 541.682705] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.682705] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] _ensure_no_port_binding_failure(port) [ 541.682705] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.682705] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] raise exception.PortBindingFailed(port_id=port['id']) [ 541.683062] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] nova.exception.PortBindingFailed: Binding failed for port 42772f00-251f-48ec-9d2a-b484bbf44126, please check neutron logs for more information. [ 541.683062] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] [ 541.683062] env[61970]: INFO nova.compute.manager [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Terminating instance [ 541.686500] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Acquiring lock "refresh_cache-0b34c674-6ef0-4c30-83e9-b4a5623ea044" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.686500] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Acquired lock "refresh_cache-0b34c674-6ef0-4c30-83e9-b4a5623ea044" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.686500] env[61970]: DEBUG nova.network.neutron [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 542.042746] env[61970]: DEBUG nova.compute.utils [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 542.046696] env[61970]: DEBUG nova.compute.manager [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Not allocating networking since 'none' was specified. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 542.229548] env[61970]: DEBUG nova.network.neutron [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.359058] env[61970]: DEBUG nova.compute.manager [req-d57c9469-abf6-4748-8430-4aee5536701b req-b5b0bd11-f906-4fb7-9137-6d1333e09849 service nova] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Received event network-vif-deleted-9644b93e-8dd4-499b-bc7b-cc741973acc2 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 542.360031] env[61970]: DEBUG nova.compute.manager [req-d57c9469-abf6-4748-8430-4aee5536701b req-b5b0bd11-f906-4fb7-9137-6d1333e09849 service nova] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Received event network-changed-42772f00-251f-48ec-9d2a-b484bbf44126 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 542.360031] env[61970]: DEBUG nova.compute.manager [req-d57c9469-abf6-4748-8430-4aee5536701b req-b5b0bd11-f906-4fb7-9137-6d1333e09849 service nova] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Refreshing instance network info cache due to event network-changed-42772f00-251f-48ec-9d2a-b484bbf44126. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 542.360031] env[61970]: DEBUG oslo_concurrency.lockutils [req-d57c9469-abf6-4748-8430-4aee5536701b req-b5b0bd11-f906-4fb7-9137-6d1333e09849 service nova] Acquiring lock "refresh_cache-0b34c674-6ef0-4c30-83e9-b4a5623ea044" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.437356] env[61970]: DEBUG nova.network.neutron [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.546945] env[61970]: DEBUG nova.compute.manager [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 542.693365] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63eafb4-fc22-497a-b440-1e3a20693d5d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.702258] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aedfb328-85eb-448f-a44f-2e25805044f0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.737133] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701ec385-9f90-4f33-a990-14edd1f2b2ef {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.746639] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2afdb03-00cf-4a2e-9bfe-daf13a5e9103 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.765481] env[61970]: DEBUG nova.compute.provider_tree [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 542.943011] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Releasing lock "refresh_cache-0b34c674-6ef0-4c30-83e9-b4a5623ea044" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.946080] env[61970]: DEBUG nova.compute.manager [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 542.946080] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 542.946080] env[61970]: DEBUG oslo_concurrency.lockutils [req-d57c9469-abf6-4748-8430-4aee5536701b req-b5b0bd11-f906-4fb7-9137-6d1333e09849 service nova] Acquired lock "refresh_cache-0b34c674-6ef0-4c30-83e9-b4a5623ea044" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.946080] env[61970]: DEBUG nova.network.neutron [req-d57c9469-abf6-4748-8430-4aee5536701b req-b5b0bd11-f906-4fb7-9137-6d1333e09849 service nova] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Refreshing network info cache for port 42772f00-251f-48ec-9d2a-b484bbf44126 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 542.947094] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e589a754-0f49-47dc-8eec-625ce0cdfc23 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.960807] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759fe9f2-4230-43d1-a08e-1bd07b9c06b7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.974636] env[61970]: DEBUG nova.compute.manager [req-b6f79e4c-51c5-4ce3-b7a9-345d808bdf2e req-f889af2e-c263-4408-a3e6-d183fb6cafb7 service nova] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Received event network-vif-deleted-6f11db79-b4bf-466c-902c-953c11ed548b {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 542.989026] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0b34c674-6ef0-4c30-83e9-b4a5623ea044 could not be found. [ 542.989026] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 542.989026] env[61970]: INFO nova.compute.manager [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Took 0.04 seconds to destroy the instance on the hypervisor. [ 542.989026] env[61970]: DEBUG oslo.service.loopingcall [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 542.989026] env[61970]: DEBUG nova.compute.manager [-] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 542.989026] env[61970]: DEBUG nova.network.neutron [-] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 543.018410] env[61970]: DEBUG nova.network.neutron [-] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.269205] env[61970]: DEBUG nova.scheduler.client.report [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 543.521156] env[61970]: DEBUG nova.network.neutron [-] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.561235] env[61970]: DEBUG nova.compute.manager [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 543.590425] env[61970]: DEBUG nova.network.neutron [req-d57c9469-abf6-4748-8430-4aee5536701b req-b5b0bd11-f906-4fb7-9137-6d1333e09849 service nova] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.596195] env[61970]: DEBUG nova.virt.hardware [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 543.596732] env[61970]: DEBUG nova.virt.hardware [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 543.597232] env[61970]: DEBUG nova.virt.hardware [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 543.598144] env[61970]: DEBUG nova.virt.hardware [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 543.598144] env[61970]: DEBUG nova.virt.hardware [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 543.598794] env[61970]: DEBUG nova.virt.hardware [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 543.598794] env[61970]: DEBUG nova.virt.hardware [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 543.599060] env[61970]: DEBUG nova.virt.hardware [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 543.599467] env[61970]: DEBUG nova.virt.hardware [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 543.599726] env[61970]: DEBUG nova.virt.hardware [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 543.600088] env[61970]: DEBUG nova.virt.hardware [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 543.602165] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c558d856-796e-44c8-a782-caf27428c0e8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.613357] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7653ea29-9d83-4258-a5fb-658e44b192c8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.635195] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Instance VIF info [] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 543.648440] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 543.649852] env[61970]: ERROR nova.compute.manager [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e544516b-8f51-45de-9462-e55eae71c56b, please check neutron logs for more information. [ 543.649852] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 543.649852] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.649852] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 543.649852] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.649852] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 543.649852] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.649852] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 543.649852] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.649852] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 543.649852] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.649852] env[61970]: ERROR nova.compute.manager raise self.value [ 543.649852] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.649852] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 543.649852] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.649852] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 543.650307] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.650307] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 543.650307] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e544516b-8f51-45de-9462-e55eae71c56b, please check neutron logs for more information. [ 543.650307] env[61970]: ERROR nova.compute.manager [ 543.650307] env[61970]: Traceback (most recent call last): [ 543.650307] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 543.650307] env[61970]: listener.cb(fileno) [ 543.650307] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.650307] env[61970]: result = function(*args, **kwargs) [ 543.650307] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 543.650307] env[61970]: return func(*args, **kwargs) [ 543.650307] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.650307] env[61970]: raise e [ 543.650307] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.650307] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 543.650307] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.650307] env[61970]: created_port_ids = self._update_ports_for_instance( [ 543.650307] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.650307] env[61970]: with excutils.save_and_reraise_exception(): [ 543.650307] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.650307] env[61970]: self.force_reraise() [ 543.650307] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.650307] env[61970]: raise self.value [ 543.650307] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.650307] env[61970]: updated_port = self._update_port( [ 543.650307] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.650307] env[61970]: _ensure_no_port_binding_failure(port) [ 543.650307] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.650307] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 543.651041] env[61970]: nova.exception.PortBindingFailed: Binding failed for port e544516b-8f51-45de-9462-e55eae71c56b, please check neutron logs for more information. [ 543.651041] env[61970]: Removing descriptor: 16 [ 543.651041] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dd312631-0e62-49c2-b948-f66825a7f19a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.653184] env[61970]: ERROR nova.compute.manager [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e544516b-8f51-45de-9462-e55eae71c56b, please check neutron logs for more information. [ 543.653184] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Traceback (most recent call last): [ 543.653184] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 543.653184] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] yield resources [ 543.653184] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 543.653184] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] self.driver.spawn(context, instance, image_meta, [ 543.653184] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 543.653184] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 543.653184] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 543.653184] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] vm_ref = self.build_virtual_machine(instance, [ 543.653184] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 543.656239] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] vif_infos = vmwarevif.get_vif_info(self._session, [ 543.656239] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 543.656239] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] for vif in network_info: [ 543.656239] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 543.656239] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] return self._sync_wrapper(fn, *args, **kwargs) [ 543.656239] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 543.656239] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] self.wait() [ 543.656239] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 543.656239] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] self[:] = self._gt.wait() [ 543.656239] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 543.656239] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] return self._exit_event.wait() [ 543.656239] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 543.656239] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] result = hub.switch() [ 543.656811] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 543.656811] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] return self.greenlet.switch() [ 543.656811] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.656811] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] result = function(*args, **kwargs) [ 543.656811] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 543.656811] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] return func(*args, **kwargs) [ 543.656811] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.656811] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] raise e [ 543.656811] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.656811] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] nwinfo = self.network_api.allocate_for_instance( [ 543.656811] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.656811] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] created_port_ids = self._update_ports_for_instance( [ 543.656811] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.657213] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] with excutils.save_and_reraise_exception(): [ 543.657213] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.657213] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] self.force_reraise() [ 543.657213] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.657213] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] raise self.value [ 543.657213] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.657213] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] updated_port = self._update_port( [ 543.657213] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.657213] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] _ensure_no_port_binding_failure(port) [ 543.657213] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.657213] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] raise exception.PortBindingFailed(port_id=port['id']) [ 543.657213] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] nova.exception.PortBindingFailed: Binding failed for port e544516b-8f51-45de-9462-e55eae71c56b, please check neutron logs for more information. [ 543.657213] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] [ 543.657565] env[61970]: INFO nova.compute.manager [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Terminating instance [ 543.657565] env[61970]: DEBUG oslo_concurrency.lockutils [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Acquiring lock "refresh_cache-4d9ee781-bca8-4cad-9fc5-87cb785d478e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.657565] env[61970]: DEBUG oslo_concurrency.lockutils [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Acquired lock "refresh_cache-4d9ee781-bca8-4cad-9fc5-87cb785d478e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.657565] env[61970]: DEBUG nova.network.neutron [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 543.667699] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Created folder: OpenStack in parent group-v4. [ 543.669656] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Creating folder: Project (8a7a27d780ba4fa0ad2f227d7648104a). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 543.669656] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-824e3ca8-1176-4a4a-a880-5a41a3a503d1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.681112] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Created folder: Project (8a7a27d780ba4fa0ad2f227d7648104a) in parent group-v288740. [ 543.681927] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Creating folder: Instances. Parent ref: group-v288741. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 543.681927] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-77be0773-f3c3-473b-ac2c-61cd94f05219 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.692626] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Created folder: Instances in parent group-v288741. [ 543.692679] env[61970]: DEBUG oslo.service.loopingcall [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 543.692947] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 543.693136] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d8c210ac-8249-486d-909f-450bdaed16e4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.715654] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 543.715654] env[61970]: value = "task-1355462" [ 543.715654] env[61970]: _type = "Task" [ 543.715654] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.725709] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355462, 'name': CreateVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.736695] env[61970]: DEBUG nova.network.neutron [req-d57c9469-abf6-4748-8430-4aee5536701b req-b5b0bd11-f906-4fb7-9137-6d1333e09849 service nova] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.774214] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.235s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.774727] env[61970]: DEBUG nova.compute.manager [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 543.779223] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.052s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.781543] env[61970]: INFO nova.compute.claims [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 543.936258] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Acquiring lock "8881014c-8505-4e75-8c31-e25b6a8d2cd4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.936258] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Lock "8881014c-8505-4e75-8c31-e25b6a8d2cd4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.025329] env[61970]: INFO nova.compute.manager [-] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Took 1.04 seconds to deallocate network for instance. [ 544.031803] env[61970]: DEBUG nova.compute.claims [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 544.032028] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.202653] env[61970]: DEBUG nova.network.neutron [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.230907] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355462, 'name': CreateVM_Task, 'duration_secs': 0.290672} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.231302] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 544.232512] env[61970]: DEBUG oslo_vmware.service [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab93e08-72c6-431b-86ee-161cbaf1a01a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.240403] env[61970]: DEBUG oslo_concurrency.lockutils [req-d57c9469-abf6-4748-8430-4aee5536701b req-b5b0bd11-f906-4fb7-9137-6d1333e09849 service nova] Releasing lock "refresh_cache-0b34c674-6ef0-4c30-83e9-b4a5623ea044" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.240403] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.240403] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.240762] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 544.241535] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-feda55b7-0800-45e9-a526-cc8cad24ff9f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.250911] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Waiting for the task: (returnval){ [ 544.250911] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5218ac2e-b96d-a662-2557-00f5c2f16a04" [ 544.250911] env[61970]: _type = "Task" [ 544.250911] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.259159] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5218ac2e-b96d-a662-2557-00f5c2f16a04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.290807] env[61970]: DEBUG nova.compute.utils [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 544.297900] env[61970]: DEBUG nova.compute.manager [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 544.297900] env[61970]: DEBUG nova.network.neutron [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 544.403033] env[61970]: DEBUG nova.network.neutron [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.431544] env[61970]: DEBUG nova.policy [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ef80257aeedb480ebfca4238d2cd7c8e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91807992e5154986ba4c6a51616397aa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 544.440338] env[61970]: DEBUG nova.compute.manager [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 544.767883] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.768137] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 544.768372] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.768514] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.768977] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 544.769215] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-94615f89-4f3c-445a-a212-3d1dfc73a553 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.785750] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 544.785931] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 544.787051] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a132d1-8afc-4e9b-bee2-8562bfff338b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.795400] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e38bc9a4-6f54-498a-8967-c6e110a560bf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.798570] env[61970]: DEBUG nova.compute.manager [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 544.806121] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Waiting for the task: (returnval){ [ 544.806121] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52240e8b-0583-25e7-f657-5304f568a9f6" [ 544.806121] env[61970]: _type = "Task" [ 544.806121] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.816364] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Preparing fetch location {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 544.816676] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Creating directory with path [datastore2] vmware_temp/612fba34-cb2e-4d7f-bf2b-2ac4d4f10e61/7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 544.817246] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9a9ec097-1659-4f2e-a8fa-b4677f1a53be {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.850882] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Created directory with path [datastore2] vmware_temp/612fba34-cb2e-4d7f-bf2b-2ac4d4f10e61/7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 544.850882] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Fetch image to [datastore2] vmware_temp/612fba34-cb2e-4d7f-bf2b-2ac4d4f10e61/7ff5af44-3f73-4c84-82e5-a1543db26d03/tmp-sparse.vmdk {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 544.850882] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Downloading image file data 7ff5af44-3f73-4c84-82e5-a1543db26d03 to [datastore2] vmware_temp/612fba34-cb2e-4d7f-bf2b-2ac4d4f10e61/7ff5af44-3f73-4c84-82e5-a1543db26d03/tmp-sparse.vmdk on the data store datastore2 {{(pid=61970) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 544.851222] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f431f5-81d8-4888-b91a-5f79b25e6a42 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.865828] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce0693e-eba0-4483-9130-40faec109597 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.878166] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21c11986-0a30-4527-a362-35846fcf9901 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.912016] env[61970]: DEBUG oslo_concurrency.lockutils [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Releasing lock "refresh_cache-4d9ee781-bca8-4cad-9fc5-87cb785d478e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.912509] env[61970]: DEBUG nova.compute.manager [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 544.912703] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 544.916022] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6e4ea3dc-283a-4257-b0fa-352b93cc0fef {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.918082] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff59c95-f89e-43d6-9887-fec9735706b3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.928832] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e23c80-b986-4e6b-b7a2-769c2e438e3a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.953264] env[61970]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-e4c66b3c-a50c-4b9e-a34c-843487dea3f0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.961250] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4d9ee781-bca8-4cad-9fc5-87cb785d478e could not be found. [ 544.961520] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 544.961704] env[61970]: INFO nova.compute.manager [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Took 0.05 seconds to destroy the instance on the hypervisor. [ 544.962058] env[61970]: DEBUG oslo.service.loopingcall [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 544.966554] env[61970]: DEBUG nova.compute.manager [-] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 544.966678] env[61970]: DEBUG nova.network.neutron [-] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 544.987794] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.993362] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Downloading image file data 7ff5af44-3f73-4c84-82e5-a1543db26d03 to the data store datastore2 {{(pid=61970) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 545.067119] env[61970]: DEBUG nova.network.neutron [-] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.088112] env[61970]: DEBUG oslo_vmware.rw_handles [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/612fba34-cb2e-4d7f-bf2b-2ac4d4f10e61/7ff5af44-3f73-4c84-82e5-a1543db26d03/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61970) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 545.088112] env[61970]: DEBUG nova.network.neutron [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Successfully created port: 3b92f861-eeb4-4364-9ab4-21f10dce1713 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 545.151860] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fe2c694-f4bf-46a8-88ae-0a4fa680dcac {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.164550] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e96ecff0-e861-4cdc-b52d-66e93dc65689 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.200377] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75a13efd-3351-4458-80f1-50b2f1e8ad4c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.209482] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83fd4b6a-b3e4-4208-aa3a-1265598893e8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.227306] env[61970]: DEBUG nova.compute.provider_tree [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 545.244674] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Acquiring lock "604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.244910] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Lock "604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.478458] env[61970]: DEBUG nova.compute.manager [req-0ab6ee01-5d72-4df4-91d4-a0f74eaff943 req-2b6f0e80-a664-4b01-90ee-e69d99bcbeff service nova] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Received event network-vif-deleted-42772f00-251f-48ec-9d2a-b484bbf44126 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 545.509837] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Acquiring lock "39bc8558-ce7f-41e9-b913-9b628a220ca7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.510209] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Lock "39bc8558-ce7f-41e9-b913-9b628a220ca7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.570095] env[61970]: DEBUG nova.network.neutron [-] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.736019] env[61970]: DEBUG nova.scheduler.client.report [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 545.751015] env[61970]: DEBUG nova.compute.manager [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 545.818301] env[61970]: DEBUG nova.compute.manager [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 545.847371] env[61970]: DEBUG oslo_vmware.rw_handles [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Completed reading data from the image iterator. {{(pid=61970) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 545.847371] env[61970]: DEBUG oslo_vmware.rw_handles [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/612fba34-cb2e-4d7f-bf2b-2ac4d4f10e61/7ff5af44-3f73-4c84-82e5-a1543db26d03/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61970) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 545.862123] env[61970]: DEBUG nova.virt.hardware [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 545.862123] env[61970]: DEBUG nova.virt.hardware [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 545.862123] env[61970]: DEBUG nova.virt.hardware [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 545.862123] env[61970]: DEBUG nova.virt.hardware [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 545.862310] env[61970]: DEBUG nova.virt.hardware [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 545.862310] env[61970]: DEBUG nova.virt.hardware [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 545.862310] env[61970]: DEBUG nova.virt.hardware [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 545.862310] env[61970]: DEBUG nova.virt.hardware [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 545.862310] env[61970]: DEBUG nova.virt.hardware [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 545.863411] env[61970]: DEBUG nova.virt.hardware [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 545.863612] env[61970]: DEBUG nova.virt.hardware [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 545.865340] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b10afe88-0a69-4d1d-91f8-a66baadc308f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.876778] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c873a38f-7fcb-4f28-b1f8-aaf8bd815291 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.884461] env[61970]: DEBUG nova.compute.manager [req-2f1eb984-b8d2-4082-bf99-87bf34e29689 req-8486c905-0bba-45e0-8e6f-81b868426bfc service nova] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Received event network-changed-e544516b-8f51-45de-9462-e55eae71c56b {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 545.884658] env[61970]: DEBUG nova.compute.manager [req-2f1eb984-b8d2-4082-bf99-87bf34e29689 req-8486c905-0bba-45e0-8e6f-81b868426bfc service nova] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Refreshing instance network info cache due to event network-changed-e544516b-8f51-45de-9462-e55eae71c56b. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 545.884896] env[61970]: DEBUG oslo_concurrency.lockutils [req-2f1eb984-b8d2-4082-bf99-87bf34e29689 req-8486c905-0bba-45e0-8e6f-81b868426bfc service nova] Acquiring lock "refresh_cache-4d9ee781-bca8-4cad-9fc5-87cb785d478e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.888036] env[61970]: DEBUG oslo_concurrency.lockutils [req-2f1eb984-b8d2-4082-bf99-87bf34e29689 req-8486c905-0bba-45e0-8e6f-81b868426bfc service nova] Acquired lock "refresh_cache-4d9ee781-bca8-4cad-9fc5-87cb785d478e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.888036] env[61970]: DEBUG nova.network.neutron [req-2f1eb984-b8d2-4082-bf99-87bf34e29689 req-8486c905-0bba-45e0-8e6f-81b868426bfc service nova] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Refreshing network info cache for port e544516b-8f51-45de-9462-e55eae71c56b {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 545.897332] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Downloaded image file data 7ff5af44-3f73-4c84-82e5-a1543db26d03 to vmware_temp/612fba34-cb2e-4d7f-bf2b-2ac4d4f10e61/7ff5af44-3f73-4c84-82e5-a1543db26d03/tmp-sparse.vmdk on the data store datastore2 {{(pid=61970) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 545.899151] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Caching image {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 545.899240] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Copying Virtual Disk [datastore2] vmware_temp/612fba34-cb2e-4d7f-bf2b-2ac4d4f10e61/7ff5af44-3f73-4c84-82e5-a1543db26d03/tmp-sparse.vmdk to [datastore2] vmware_temp/612fba34-cb2e-4d7f-bf2b-2ac4d4f10e61/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 545.903106] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c83ee1de-0c2b-4652-a980-918355b392e4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.911602] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Waiting for the task: (returnval){ [ 545.911602] env[61970]: value = "task-1355463" [ 545.911602] env[61970]: _type = "Task" [ 545.911602] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.922727] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355463, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.959281] env[61970]: DEBUG nova.network.neutron [req-2f1eb984-b8d2-4082-bf99-87bf34e29689 req-8486c905-0bba-45e0-8e6f-81b868426bfc service nova] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.073238] env[61970]: INFO nova.compute.manager [-] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Took 1.11 seconds to deallocate network for instance. [ 546.075921] env[61970]: DEBUG nova.compute.claims [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 546.076179] env[61970]: DEBUG oslo_concurrency.lockutils [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.227403] env[61970]: DEBUG nova.network.neutron [req-2f1eb984-b8d2-4082-bf99-87bf34e29689 req-8486c905-0bba-45e0-8e6f-81b868426bfc service nova] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.239316] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.459s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.239316] env[61970]: DEBUG nova.compute.manager [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 546.241478] env[61970]: DEBUG oslo_concurrency.lockutils [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.324s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.243498] env[61970]: INFO nova.compute.claims [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 546.284502] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.424845] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355463, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.734722] env[61970]: DEBUG oslo_concurrency.lockutils [req-2f1eb984-b8d2-4082-bf99-87bf34e29689 req-8486c905-0bba-45e0-8e6f-81b868426bfc service nova] Releasing lock "refresh_cache-4d9ee781-bca8-4cad-9fc5-87cb785d478e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.734963] env[61970]: DEBUG nova.compute.manager [req-2f1eb984-b8d2-4082-bf99-87bf34e29689 req-8486c905-0bba-45e0-8e6f-81b868426bfc service nova] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Received event network-vif-deleted-e544516b-8f51-45de-9462-e55eae71c56b {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 546.747902] env[61970]: DEBUG nova.compute.utils [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 546.756730] env[61970]: DEBUG nova.compute.manager [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 546.756730] env[61970]: DEBUG nova.network.neutron [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 546.925793] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355463, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.693419} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.925948] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Copied Virtual Disk [datastore2] vmware_temp/612fba34-cb2e-4d7f-bf2b-2ac4d4f10e61/7ff5af44-3f73-4c84-82e5-a1543db26d03/tmp-sparse.vmdk to [datastore2] vmware_temp/612fba34-cb2e-4d7f-bf2b-2ac4d4f10e61/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 546.927919] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Deleting the datastore file [datastore2] vmware_temp/612fba34-cb2e-4d7f-bf2b-2ac4d4f10e61/7ff5af44-3f73-4c84-82e5-a1543db26d03/tmp-sparse.vmdk {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 546.927919] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-66f85242-cd7a-4b3f-89c9-11ca659a0b60 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.935497] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Waiting for the task: (returnval){ [ 546.935497] env[61970]: value = "task-1355464" [ 546.935497] env[61970]: _type = "Task" [ 546.935497] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.948944] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355464, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.970068] env[61970]: DEBUG nova.policy [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '57c3e42bcf68420e9d3791f89ae1589f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb20fc25493f42529b1538b73f5a2938', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 546.990607] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Acquiring lock "17f41a87-4fd0-4866-905a-2aa3d2e9691b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.990854] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Lock "17f41a87-4fd0-4866-905a-2aa3d2e9691b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.266911] env[61970]: DEBUG nova.compute.manager [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 547.448607] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355464, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023765} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.451634] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 547.451786] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Moving file from [datastore2] vmware_temp/612fba34-cb2e-4d7f-bf2b-2ac4d4f10e61/7ff5af44-3f73-4c84-82e5-a1543db26d03 to [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03. {{(pid=61970) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 547.452728] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-c3eb2c10-258b-4b83-bf1e-b4431a2eb4c6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.463480] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Waiting for the task: (returnval){ [ 547.463480] env[61970]: value = "task-1355465" [ 547.463480] env[61970]: _type = "Task" [ 547.463480] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.475296] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355465, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.484401] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c4c364-1d44-4ea3-a681-aa5bf9ea9dd8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.492277] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f03cfe98-2ea4-4606-96c1-2c63e9509dcb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.526744] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a08ca55a-f95b-43a0-b915-ece889590f3f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.534372] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf836f9b-d452-4e1d-a2d5-1552de508833 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.551439] env[61970]: DEBUG nova.compute.provider_tree [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.800519] env[61970]: DEBUG oslo_concurrency.lockutils [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Acquiring lock "8bacc19d-2626-406a-a3fb-f1069674963c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.800861] env[61970]: DEBUG oslo_concurrency.lockutils [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Lock "8bacc19d-2626-406a-a3fb-f1069674963c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.975562] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355465, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.027774} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.975870] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] File moved {{(pid=61970) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 547.977276] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Cleaning up location [datastore2] vmware_temp/612fba34-cb2e-4d7f-bf2b-2ac4d4f10e61 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 547.977499] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Deleting the datastore file [datastore2] vmware_temp/612fba34-cb2e-4d7f-bf2b-2ac4d4f10e61 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 547.982053] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d1fed74c-4e7d-4809-af8c-bfa341b35096 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.986580] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Waiting for the task: (returnval){ [ 547.986580] env[61970]: value = "task-1355466" [ 547.986580] env[61970]: _type = "Task" [ 547.986580] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.999970] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355466, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.056286] env[61970]: DEBUG nova.scheduler.client.report [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 548.281614] env[61970]: DEBUG nova.compute.manager [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 548.310809] env[61970]: DEBUG nova.virt.hardware [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 548.311323] env[61970]: DEBUG nova.virt.hardware [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 548.311637] env[61970]: DEBUG nova.virt.hardware [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 548.311978] env[61970]: DEBUG nova.virt.hardware [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 548.316017] env[61970]: DEBUG nova.virt.hardware [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 548.316017] env[61970]: DEBUG nova.virt.hardware [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 548.316017] env[61970]: DEBUG nova.virt.hardware [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 548.316017] env[61970]: DEBUG nova.virt.hardware [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 548.316017] env[61970]: DEBUG nova.virt.hardware [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 548.316642] env[61970]: DEBUG nova.virt.hardware [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 548.316642] env[61970]: DEBUG nova.virt.hardware [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 548.316642] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524423e1-4855-4f58-8ac7-7cfc49c17d21 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.326598] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5260eea1-2c7c-4bcf-84cf-7d723b66286a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.478883] env[61970]: DEBUG nova.network.neutron [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Successfully created port: aae58430-6a65-4482-befa-f615cc7055d5 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 548.499253] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355466, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025462} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.499649] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 548.500602] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5ac487a-4cfa-424c-8ee8-c121653ed707 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.507774] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Waiting for the task: (returnval){ [ 548.507774] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d52003-1dd3-dc50-a455-84770bf834fd" [ 548.507774] env[61970]: _type = "Task" [ 548.507774] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.519311] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d52003-1dd3-dc50-a455-84770bf834fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.544110] env[61970]: ERROR nova.compute.manager [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3b92f861-eeb4-4364-9ab4-21f10dce1713, please check neutron logs for more information. [ 548.544110] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 548.544110] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.544110] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 548.544110] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 548.544110] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 548.544110] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 548.544110] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 548.544110] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.544110] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 548.544110] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.544110] env[61970]: ERROR nova.compute.manager raise self.value [ 548.544110] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 548.544110] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 548.544110] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.544110] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 548.544616] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.544616] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 548.544616] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3b92f861-eeb4-4364-9ab4-21f10dce1713, please check neutron logs for more information. [ 548.544616] env[61970]: ERROR nova.compute.manager [ 548.544616] env[61970]: Traceback (most recent call last): [ 548.544616] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 548.544616] env[61970]: listener.cb(fileno) [ 548.544616] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.544616] env[61970]: result = function(*args, **kwargs) [ 548.544616] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 548.544616] env[61970]: return func(*args, **kwargs) [ 548.544616] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 548.544616] env[61970]: raise e [ 548.544616] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.544616] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 548.544616] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 548.544616] env[61970]: created_port_ids = self._update_ports_for_instance( [ 548.544616] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 548.544616] env[61970]: with excutils.save_and_reraise_exception(): [ 548.544616] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.544616] env[61970]: self.force_reraise() [ 548.544616] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.544616] env[61970]: raise self.value [ 548.544616] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 548.544616] env[61970]: updated_port = self._update_port( [ 548.544616] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.544616] env[61970]: _ensure_no_port_binding_failure(port) [ 548.544616] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.544616] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 548.545591] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 3b92f861-eeb4-4364-9ab4-21f10dce1713, please check neutron logs for more information. [ 548.545591] env[61970]: Removing descriptor: 16 [ 548.545591] env[61970]: ERROR nova.compute.manager [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3b92f861-eeb4-4364-9ab4-21f10dce1713, please check neutron logs for more information. [ 548.545591] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Traceback (most recent call last): [ 548.545591] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 548.545591] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] yield resources [ 548.545591] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 548.545591] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] self.driver.spawn(context, instance, image_meta, [ 548.545591] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 548.545591] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 548.545591] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 548.545591] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] vm_ref = self.build_virtual_machine(instance, [ 548.545914] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 548.545914] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] vif_infos = vmwarevif.get_vif_info(self._session, [ 548.545914] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 548.545914] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] for vif in network_info: [ 548.545914] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 548.545914] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] return self._sync_wrapper(fn, *args, **kwargs) [ 548.545914] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 548.545914] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] self.wait() [ 548.545914] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 548.545914] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] self[:] = self._gt.wait() [ 548.545914] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 548.545914] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] return self._exit_event.wait() [ 548.545914] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 548.546305] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] result = hub.switch() [ 548.546305] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 548.546305] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] return self.greenlet.switch() [ 548.546305] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.546305] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] result = function(*args, **kwargs) [ 548.546305] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 548.546305] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] return func(*args, **kwargs) [ 548.546305] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 548.546305] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] raise e [ 548.546305] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.546305] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] nwinfo = self.network_api.allocate_for_instance( [ 548.546305] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 548.546305] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] created_port_ids = self._update_ports_for_instance( [ 548.546654] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 548.546654] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] with excutils.save_and_reraise_exception(): [ 548.546654] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.546654] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] self.force_reraise() [ 548.546654] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.546654] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] raise self.value [ 548.546654] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 548.546654] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] updated_port = self._update_port( [ 548.546654] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.546654] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] _ensure_no_port_binding_failure(port) [ 548.546654] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.546654] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] raise exception.PortBindingFailed(port_id=port['id']) [ 548.546956] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] nova.exception.PortBindingFailed: Binding failed for port 3b92f861-eeb4-4364-9ab4-21f10dce1713, please check neutron logs for more information. [ 548.546956] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] [ 548.546956] env[61970]: INFO nova.compute.manager [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Terminating instance [ 548.551919] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Acquiring lock "refresh_cache-bce7d001-2885-409c-b455-4c8a3ce8eb9e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.552094] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Acquired lock "refresh_cache-bce7d001-2885-409c-b455-4c8a3ce8eb9e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.552261] env[61970]: DEBUG nova.network.neutron [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 548.569044] env[61970]: DEBUG oslo_concurrency.lockutils [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.327s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.569510] env[61970]: DEBUG nova.compute.manager [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 548.572815] env[61970]: DEBUG oslo_concurrency.lockutils [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.370s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.024031] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d52003-1dd3-dc50-a455-84770bf834fd, 'name': SearchDatastore_Task, 'duration_secs': 0.009217} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.024752] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.026638] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] c4d3b145-84be-4d19-afc7-4c29fa193760/c4d3b145-84be-4d19-afc7-4c29fa193760.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 549.026638] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-400a7c19-23ce-4ef4-b1ba-5277f203d35b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.038226] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Waiting for the task: (returnval){ [ 549.038226] env[61970]: value = "task-1355467" [ 549.038226] env[61970]: _type = "Task" [ 549.038226] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.047426] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355467, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.077143] env[61970]: DEBUG nova.compute.utils [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 549.082872] env[61970]: DEBUG nova.compute.manager [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 549.083598] env[61970]: DEBUG nova.network.neutron [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 549.197602] env[61970]: DEBUG nova.network.neutron [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.301784] env[61970]: DEBUG nova.policy [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '92bef65b60c245c18531607c9d1b65d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff5a47aa0888479db5613dfd4bf53fca', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 549.349409] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d28083-7a15-4db2-8fb9-e1acfded2f74 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.364033] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a819efe-1973-40fa-b170-a1129577ad8a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.404160] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8daa0ce2-b7c3-47a2-88f0-2c5994e55fbe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.415136] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb0577e-4ad7-428b-aead-3e15dd7e78a5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.429557] env[61970]: DEBUG nova.compute.provider_tree [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 549.558742] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355467, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.587267] env[61970]: DEBUG nova.compute.manager [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 549.799090] env[61970]: DEBUG nova.network.neutron [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.938238] env[61970]: DEBUG nova.scheduler.client.report [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 550.046874] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355467, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.587391} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.047173] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] c4d3b145-84be-4d19-afc7-4c29fa193760/c4d3b145-84be-4d19-afc7-4c29fa193760.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 550.047405] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 550.047674] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-65130d94-a373-4365-8f05-36a56d594c96 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.056924] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Waiting for the task: (returnval){ [ 550.056924] env[61970]: value = "task-1355468" [ 550.056924] env[61970]: _type = "Task" [ 550.056924] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.068558] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355468, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.302121] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Releasing lock "refresh_cache-bce7d001-2885-409c-b455-4c8a3ce8eb9e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.302565] env[61970]: DEBUG nova.compute.manager [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 550.302743] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 550.303744] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a6bd70f2-af27-40c4-a982-582d55f0ca91 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.316688] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3afed1bd-1ca5-42c0-a59e-be753c8ab7f0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.346550] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bce7d001-2885-409c-b455-4c8a3ce8eb9e could not be found. [ 550.346550] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 550.346550] env[61970]: INFO nova.compute.manager [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 550.346550] env[61970]: DEBUG oslo.service.loopingcall [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 550.346550] env[61970]: DEBUG nova.compute.manager [-] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 550.346550] env[61970]: DEBUG nova.network.neutron [-] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 550.396980] env[61970]: DEBUG nova.network.neutron [-] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.417864] env[61970]: DEBUG nova.compute.manager [req-8f105212-794e-4f49-903e-c79b1cafc326 req-1ebdf9cd-63da-4165-aea3-5ae21eb3a260 service nova] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Received event network-changed-3b92f861-eeb4-4364-9ab4-21f10dce1713 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 550.418058] env[61970]: DEBUG nova.compute.manager [req-8f105212-794e-4f49-903e-c79b1cafc326 req-1ebdf9cd-63da-4165-aea3-5ae21eb3a260 service nova] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Refreshing instance network info cache due to event network-changed-3b92f861-eeb4-4364-9ab4-21f10dce1713. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 550.418290] env[61970]: DEBUG oslo_concurrency.lockutils [req-8f105212-794e-4f49-903e-c79b1cafc326 req-1ebdf9cd-63da-4165-aea3-5ae21eb3a260 service nova] Acquiring lock "refresh_cache-bce7d001-2885-409c-b455-4c8a3ce8eb9e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.418579] env[61970]: DEBUG oslo_concurrency.lockutils [req-8f105212-794e-4f49-903e-c79b1cafc326 req-1ebdf9cd-63da-4165-aea3-5ae21eb3a260 service nova] Acquired lock "refresh_cache-bce7d001-2885-409c-b455-4c8a3ce8eb9e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.418659] env[61970]: DEBUG nova.network.neutron [req-8f105212-794e-4f49-903e-c79b1cafc326 req-1ebdf9cd-63da-4165-aea3-5ae21eb3a260 service nova] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Refreshing network info cache for port 3b92f861-eeb4-4364-9ab4-21f10dce1713 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 550.448916] env[61970]: DEBUG oslo_concurrency.lockutils [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.876s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.449607] env[61970]: ERROR nova.compute.manager [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9644b93e-8dd4-499b-bc7b-cc741973acc2, please check neutron logs for more information. [ 550.449607] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Traceback (most recent call last): [ 550.449607] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 550.449607] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] self.driver.spawn(context, instance, image_meta, [ 550.449607] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 550.449607] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 550.449607] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 550.449607] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] vm_ref = self.build_virtual_machine(instance, [ 550.449607] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 550.449607] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] vif_infos = vmwarevif.get_vif_info(self._session, [ 550.449607] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 550.449976] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] for vif in network_info: [ 550.449976] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 550.449976] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] return self._sync_wrapper(fn, *args, **kwargs) [ 550.449976] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 550.449976] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] self.wait() [ 550.449976] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 550.449976] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] self[:] = self._gt.wait() [ 550.449976] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 550.449976] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] return self._exit_event.wait() [ 550.449976] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 550.449976] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] result = hub.switch() [ 550.449976] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 550.449976] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] return self.greenlet.switch() [ 550.450414] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 550.450414] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] result = function(*args, **kwargs) [ 550.450414] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 550.450414] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] return func(*args, **kwargs) [ 550.450414] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 550.450414] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] raise e [ 550.450414] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 550.450414] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] nwinfo = self.network_api.allocate_for_instance( [ 550.450414] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 550.450414] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] created_port_ids = self._update_ports_for_instance( [ 550.450414] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 550.450414] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] with excutils.save_and_reraise_exception(): [ 550.450414] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.450774] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] self.force_reraise() [ 550.450774] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.450774] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] raise self.value [ 550.450774] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 550.450774] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] updated_port = self._update_port( [ 550.450774] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.450774] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] _ensure_no_port_binding_failure(port) [ 550.450774] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.450774] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] raise exception.PortBindingFailed(port_id=port['id']) [ 550.450774] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] nova.exception.PortBindingFailed: Binding failed for port 9644b93e-8dd4-499b-bc7b-cc741973acc2, please check neutron logs for more information. [ 550.450774] env[61970]: ERROR nova.compute.manager [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] [ 550.451078] env[61970]: DEBUG nova.compute.utils [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Binding failed for port 9644b93e-8dd4-499b-bc7b-cc741973acc2, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 550.452021] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 10.158s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.452021] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.452021] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61970) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 550.452021] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.688s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.456979] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa0dc410-7ece-42e7-a4d6-2caf0e07e54f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.469282] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaeac12f-5721-4791-bb60-64badef16f1d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.477062] env[61970]: DEBUG nova.compute.manager [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Build of instance 3ff41f60-357c-4b56-b106-0e69207b78a6 was re-scheduled: Binding failed for port 9644b93e-8dd4-499b-bc7b-cc741973acc2, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 550.477560] env[61970]: DEBUG nova.compute.manager [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 550.477789] env[61970]: DEBUG oslo_concurrency.lockutils [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Acquiring lock "refresh_cache-3ff41f60-357c-4b56-b106-0e69207b78a6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.477930] env[61970]: DEBUG oslo_concurrency.lockutils [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Acquired lock "refresh_cache-3ff41f60-357c-4b56-b106-0e69207b78a6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.478102] env[61970]: DEBUG nova.network.neutron [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 550.499733] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16990a3-c3e1-4305-a5fc-2280b14b0506 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.512389] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Acquiring lock "fb0aec23-e010-4f9c-8a1d-1b31d17190b8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.512771] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Lock "fb0aec23-e010-4f9c-8a1d-1b31d17190b8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.519161] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7492cf30-9335-49c1-a023-87b7442c3a47 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.553110] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181529MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61970) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 550.553110] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.569646] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355468, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063585} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.569646] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 550.570328] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b88c9e-21d1-4191-a39d-b6e6abfbe0b5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.601449] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Reconfiguring VM instance instance-00000005 to attach disk [datastore2] c4d3b145-84be-4d19-afc7-4c29fa193760/c4d3b145-84be-4d19-afc7-4c29fa193760.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 550.601449] env[61970]: DEBUG nova.compute.manager [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 550.606807] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34f0e0a1-5a2e-4d80-8b3c-f048bd6bda42 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.633156] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Waiting for the task: (returnval){ [ 550.633156] env[61970]: value = "task-1355469" [ 550.633156] env[61970]: _type = "Task" [ 550.633156] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.647216] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355469, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.664584] env[61970]: DEBUG nova.virt.hardware [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 550.667433] env[61970]: DEBUG nova.virt.hardware [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 550.667433] env[61970]: DEBUG nova.virt.hardware [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 550.667433] env[61970]: DEBUG nova.virt.hardware [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 550.667433] env[61970]: DEBUG nova.virt.hardware [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 550.667433] env[61970]: DEBUG nova.virt.hardware [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 550.667739] env[61970]: DEBUG nova.virt.hardware [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 550.667739] env[61970]: DEBUG nova.virt.hardware [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 550.667739] env[61970]: DEBUG nova.virt.hardware [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 550.667739] env[61970]: DEBUG nova.virt.hardware [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 550.667739] env[61970]: DEBUG nova.virt.hardware [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 550.667946] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb95501-c811-486d-8aea-12ecc8bfba93 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.677120] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42af6884-572e-4f4a-9149-c1cd49c0b386 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.902054] env[61970]: DEBUG nova.network.neutron [-] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.005360] env[61970]: DEBUG nova.network.neutron [req-8f105212-794e-4f49-903e-c79b1cafc326 req-1ebdf9cd-63da-4165-aea3-5ae21eb3a260 service nova] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.060880] env[61970]: DEBUG nova.network.neutron [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.126964] env[61970]: DEBUG nova.network.neutron [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Successfully created port: c09b9037-d99d-4c4e-950b-27276f22b702 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 551.155019] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355469, 'name': ReconfigVM_Task, 'duration_secs': 0.280046} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.155019] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Reconfigured VM instance instance-00000005 to attach disk [datastore2] c4d3b145-84be-4d19-afc7-4c29fa193760/c4d3b145-84be-4d19-afc7-4c29fa193760.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 551.155019] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e11bca9c-51e5-4294-84dd-b346c4d8275d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.161278] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Waiting for the task: (returnval){ [ 551.161278] env[61970]: value = "task-1355470" [ 551.161278] env[61970]: _type = "Task" [ 551.161278] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.172994] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355470, 'name': Rename_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.243735] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-835f1910-7a77-4f6a-bceb-a2ffd58668eb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.252227] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506633fa-c36b-4bef-b72f-77fbd834c9bf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.288506] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e304b1d-ad2c-4f93-b041-c71d27467c33 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.296217] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61ef277d-63d6-4a77-846c-fed4ae7821bd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.313517] env[61970]: DEBUG nova.compute.provider_tree [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 551.329711] env[61970]: DEBUG nova.network.neutron [req-8f105212-794e-4f49-903e-c79b1cafc326 req-1ebdf9cd-63da-4165-aea3-5ae21eb3a260 service nova] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.389228] env[61970]: DEBUG nova.network.neutron [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.404727] env[61970]: INFO nova.compute.manager [-] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Took 1.06 seconds to deallocate network for instance. [ 551.406681] env[61970]: DEBUG nova.compute.claims [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 551.406862] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.674899] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355470, 'name': Rename_Task, 'duration_secs': 0.120998} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.675185] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 551.675521] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7313fc20-ae1b-480f-9cad-2dd0b2631d36 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.681986] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Waiting for the task: (returnval){ [ 551.681986] env[61970]: value = "task-1355471" [ 551.681986] env[61970]: _type = "Task" [ 551.681986] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.694825] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355471, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.817538] env[61970]: DEBUG nova.scheduler.client.report [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 551.833335] env[61970]: DEBUG oslo_concurrency.lockutils [req-8f105212-794e-4f49-903e-c79b1cafc326 req-1ebdf9cd-63da-4165-aea3-5ae21eb3a260 service nova] Releasing lock "refresh_cache-bce7d001-2885-409c-b455-4c8a3ce8eb9e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.892475] env[61970]: DEBUG oslo_concurrency.lockutils [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Releasing lock "refresh_cache-3ff41f60-357c-4b56-b106-0e69207b78a6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.892475] env[61970]: DEBUG nova.compute.manager [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 551.892606] env[61970]: DEBUG nova.compute.manager [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 551.892742] env[61970]: DEBUG nova.network.neutron [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 551.991837] env[61970]: DEBUG nova.network.neutron [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.196213] env[61970]: DEBUG oslo_vmware.api [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355471, 'name': PowerOnVM_Task, 'duration_secs': 0.450545} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.196621] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 552.196837] env[61970]: INFO nova.compute.manager [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Took 8.64 seconds to spawn the instance on the hypervisor. [ 552.197352] env[61970]: DEBUG nova.compute.manager [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 552.203368] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ff2a52-02c4-411b-a6dc-ab79d54057c0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.325411] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.871s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.325411] env[61970]: ERROR nova.compute.manager [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6f11db79-b4bf-466c-902c-953c11ed548b, please check neutron logs for more information. [ 552.325411] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Traceback (most recent call last): [ 552.325411] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 552.325411] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] self.driver.spawn(context, instance, image_meta, [ 552.325411] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 552.325411] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 552.325411] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 552.325411] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] vm_ref = self.build_virtual_machine(instance, [ 552.325768] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 552.325768] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] vif_infos = vmwarevif.get_vif_info(self._session, [ 552.325768] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 552.325768] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] for vif in network_info: [ 552.325768] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 552.325768] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] return self._sync_wrapper(fn, *args, **kwargs) [ 552.325768] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 552.325768] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] self.wait() [ 552.325768] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 552.325768] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] self[:] = self._gt.wait() [ 552.325768] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 552.325768] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] return self._exit_event.wait() [ 552.325768] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 552.326209] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] result = hub.switch() [ 552.326209] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 552.326209] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] return self.greenlet.switch() [ 552.326209] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 552.326209] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] result = function(*args, **kwargs) [ 552.326209] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 552.326209] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] return func(*args, **kwargs) [ 552.326209] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 552.326209] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] raise e [ 552.326209] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 552.326209] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] nwinfo = self.network_api.allocate_for_instance( [ 552.326209] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 552.326209] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] created_port_ids = self._update_ports_for_instance( [ 552.326573] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 552.326573] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] with excutils.save_and_reraise_exception(): [ 552.326573] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 552.326573] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] self.force_reraise() [ 552.326573] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 552.326573] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] raise self.value [ 552.326573] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 552.326573] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] updated_port = self._update_port( [ 552.326573] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 552.326573] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] _ensure_no_port_binding_failure(port) [ 552.326573] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 552.326573] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] raise exception.PortBindingFailed(port_id=port['id']) [ 552.326886] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] nova.exception.PortBindingFailed: Binding failed for port 6f11db79-b4bf-466c-902c-953c11ed548b, please check neutron logs for more information. [ 552.326886] env[61970]: ERROR nova.compute.manager [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] [ 552.326886] env[61970]: DEBUG nova.compute.utils [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Binding failed for port 6f11db79-b4bf-466c-902c-953c11ed548b, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 552.327872] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.296s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.331184] env[61970]: DEBUG nova.compute.manager [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Build of instance 10e1500f-70f1-40b6-aad6-5f9d75e1554b was re-scheduled: Binding failed for port 6f11db79-b4bf-466c-902c-953c11ed548b, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 552.331897] env[61970]: DEBUG nova.compute.manager [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 552.332236] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "refresh_cache-10e1500f-70f1-40b6-aad6-5f9d75e1554b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.332475] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquired lock "refresh_cache-10e1500f-70f1-40b6-aad6-5f9d75e1554b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.332730] env[61970]: DEBUG nova.network.neutron [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 552.496688] env[61970]: DEBUG nova.network.neutron [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.729994] env[61970]: INFO nova.compute.manager [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Took 18.49 seconds to build instance. [ 552.918447] env[61970]: DEBUG nova.network.neutron [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.001211] env[61970]: INFO nova.compute.manager [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 3ff41f60-357c-4b56-b106-0e69207b78a6] Took 1.11 seconds to deallocate network for instance. [ 553.077314] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c19a80d-0180-4d80-bf64-1fe0fb50b6c2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.089306] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de5d1cf4-bae7-486e-8a4b-021c86eaf464 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.126265] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36330f38-28cc-457d-8453-6e871bfc7aa5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.137008] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af8a42d8-f8e2-45e3-bd3f-7bec251e6bfc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.150881] env[61970]: DEBUG nova.compute.provider_tree [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 553.232977] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c0541dfe-22f3-42c9-9dad-d6e767da5c17 tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Lock "c4d3b145-84be-4d19-afc7-4c29fa193760" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.620846] env[61970]: DEBUG nova.network.neutron [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.654136] env[61970]: DEBUG nova.scheduler.client.report [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 553.702598] env[61970]: DEBUG nova.compute.manager [req-2774a0bf-f1c1-4776-bde8-bb1912b6aab0 req-6451d2cf-713d-49df-a8ee-bccc806d18a3 service nova] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Received event network-vif-deleted-3b92f861-eeb4-4364-9ab4-21f10dce1713 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 553.736542] env[61970]: DEBUG nova.compute.manager [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 554.012253] env[61970]: ERROR nova.compute.manager [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port aae58430-6a65-4482-befa-f615cc7055d5, please check neutron logs for more information. [ 554.012253] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 554.012253] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.012253] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 554.012253] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 554.012253] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 554.012253] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 554.012253] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 554.012253] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.012253] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 554.012253] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.012253] env[61970]: ERROR nova.compute.manager raise self.value [ 554.012253] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 554.012253] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 554.012253] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.012253] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 554.012779] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.012779] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 554.012779] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port aae58430-6a65-4482-befa-f615cc7055d5, please check neutron logs for more information. [ 554.012779] env[61970]: ERROR nova.compute.manager [ 554.012779] env[61970]: Traceback (most recent call last): [ 554.012779] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 554.012779] env[61970]: listener.cb(fileno) [ 554.012779] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.012779] env[61970]: result = function(*args, **kwargs) [ 554.012779] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 554.012779] env[61970]: return func(*args, **kwargs) [ 554.012779] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 554.012779] env[61970]: raise e [ 554.012779] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.012779] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 554.012779] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 554.012779] env[61970]: created_port_ids = self._update_ports_for_instance( [ 554.012779] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 554.012779] env[61970]: with excutils.save_and_reraise_exception(): [ 554.012779] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.012779] env[61970]: self.force_reraise() [ 554.012779] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.012779] env[61970]: raise self.value [ 554.012779] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 554.012779] env[61970]: updated_port = self._update_port( [ 554.012779] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.012779] env[61970]: _ensure_no_port_binding_failure(port) [ 554.012779] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.012779] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 554.013593] env[61970]: nova.exception.PortBindingFailed: Binding failed for port aae58430-6a65-4482-befa-f615cc7055d5, please check neutron logs for more information. [ 554.013593] env[61970]: Removing descriptor: 17 [ 554.018835] env[61970]: ERROR nova.compute.manager [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port aae58430-6a65-4482-befa-f615cc7055d5, please check neutron logs for more information. [ 554.018835] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Traceback (most recent call last): [ 554.018835] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 554.018835] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] yield resources [ 554.018835] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 554.018835] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] self.driver.spawn(context, instance, image_meta, [ 554.018835] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 554.018835] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 554.018835] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 554.018835] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] vm_ref = self.build_virtual_machine(instance, [ 554.018835] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 554.019277] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] vif_infos = vmwarevif.get_vif_info(self._session, [ 554.019277] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 554.019277] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] for vif in network_info: [ 554.019277] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 554.019277] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] return self._sync_wrapper(fn, *args, **kwargs) [ 554.019277] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 554.019277] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] self.wait() [ 554.019277] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 554.019277] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] self[:] = self._gt.wait() [ 554.019277] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 554.019277] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] return self._exit_event.wait() [ 554.019277] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 554.019277] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] result = hub.switch() [ 554.019737] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 554.019737] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] return self.greenlet.switch() [ 554.019737] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.019737] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] result = function(*args, **kwargs) [ 554.019737] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 554.019737] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] return func(*args, **kwargs) [ 554.019737] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 554.019737] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] raise e [ 554.019737] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.019737] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] nwinfo = self.network_api.allocate_for_instance( [ 554.019737] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 554.019737] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] created_port_ids = self._update_ports_for_instance( [ 554.019737] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 554.020164] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] with excutils.save_and_reraise_exception(): [ 554.020164] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.020164] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] self.force_reraise() [ 554.020164] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.020164] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] raise self.value [ 554.020164] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 554.020164] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] updated_port = self._update_port( [ 554.020164] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.020164] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] _ensure_no_port_binding_failure(port) [ 554.020164] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.020164] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] raise exception.PortBindingFailed(port_id=port['id']) [ 554.020164] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] nova.exception.PortBindingFailed: Binding failed for port aae58430-6a65-4482-befa-f615cc7055d5, please check neutron logs for more information. [ 554.020164] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] [ 554.020519] env[61970]: INFO nova.compute.manager [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Terminating instance [ 554.022904] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Acquiring lock "refresh_cache-36c69772-6e47-4c00-a582-b46d233e13a1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.022904] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Acquired lock "refresh_cache-36c69772-6e47-4c00-a582-b46d233e13a1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.022904] env[61970]: DEBUG nova.network.neutron [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 554.054294] env[61970]: INFO nova.scheduler.client.report [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Deleted allocations for instance 3ff41f60-357c-4b56-b106-0e69207b78a6 [ 554.125750] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Releasing lock "refresh_cache-10e1500f-70f1-40b6-aad6-5f9d75e1554b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.126522] env[61970]: DEBUG nova.compute.manager [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 554.126522] env[61970]: DEBUG nova.compute.manager [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 554.126652] env[61970]: DEBUG nova.network.neutron [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 554.163285] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.835s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.163761] env[61970]: ERROR nova.compute.manager [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 42772f00-251f-48ec-9d2a-b484bbf44126, please check neutron logs for more information. [ 554.163761] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Traceback (most recent call last): [ 554.163761] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 554.163761] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] self.driver.spawn(context, instance, image_meta, [ 554.163761] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 554.163761] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] self._vmops.spawn(context, instance, image_meta, injected_files, [ 554.163761] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 554.163761] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] vm_ref = self.build_virtual_machine(instance, [ 554.163761] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 554.163761] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] vif_infos = vmwarevif.get_vif_info(self._session, [ 554.163761] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 554.164486] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] for vif in network_info: [ 554.164486] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 554.164486] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] return self._sync_wrapper(fn, *args, **kwargs) [ 554.164486] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 554.164486] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] self.wait() [ 554.164486] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 554.164486] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] self[:] = self._gt.wait() [ 554.164486] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 554.164486] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] return self._exit_event.wait() [ 554.164486] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 554.164486] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] result = hub.switch() [ 554.164486] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 554.164486] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] return self.greenlet.switch() [ 554.164846] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.164846] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] result = function(*args, **kwargs) [ 554.164846] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 554.164846] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] return func(*args, **kwargs) [ 554.164846] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 554.164846] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] raise e [ 554.164846] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.164846] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] nwinfo = self.network_api.allocate_for_instance( [ 554.164846] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 554.164846] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] created_port_ids = self._update_ports_for_instance( [ 554.164846] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 554.164846] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] with excutils.save_and_reraise_exception(): [ 554.164846] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.165170] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] self.force_reraise() [ 554.165170] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.165170] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] raise self.value [ 554.165170] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 554.165170] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] updated_port = self._update_port( [ 554.165170] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.165170] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] _ensure_no_port_binding_failure(port) [ 554.165170] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.165170] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] raise exception.PortBindingFailed(port_id=port['id']) [ 554.165170] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] nova.exception.PortBindingFailed: Binding failed for port 42772f00-251f-48ec-9d2a-b484bbf44126, please check neutron logs for more information. [ 554.165170] env[61970]: ERROR nova.compute.manager [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] [ 554.165479] env[61970]: DEBUG nova.compute.utils [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Binding failed for port 42772f00-251f-48ec-9d2a-b484bbf44126, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 554.166935] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.179s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.169064] env[61970]: INFO nova.compute.claims [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 554.173924] env[61970]: DEBUG nova.compute.manager [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Build of instance 0b34c674-6ef0-4c30-83e9-b4a5623ea044 was re-scheduled: Binding failed for port 42772f00-251f-48ec-9d2a-b484bbf44126, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 554.174768] env[61970]: DEBUG nova.compute.manager [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 554.174816] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Acquiring lock "refresh_cache-0b34c674-6ef0-4c30-83e9-b4a5623ea044" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.174955] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Acquired lock "refresh_cache-0b34c674-6ef0-4c30-83e9-b4a5623ea044" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.175118] env[61970]: DEBUG nova.network.neutron [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 554.223360] env[61970]: DEBUG nova.network.neutron [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.265048] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.565890] env[61970]: DEBUG oslo_concurrency.lockutils [None req-25026692-0193-4b2d-a459-4a3846fdec1a tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Lock "3ff41f60-357c-4b56-b106-0e69207b78a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.346s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.582209] env[61970]: DEBUG nova.compute.manager [req-fcaac219-1160-481d-ad2a-c967af33e601 req-03ce0079-6eb1-4783-b4b0-76bada62889a service nova] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Received event network-changed-aae58430-6a65-4482-befa-f615cc7055d5 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 554.582541] env[61970]: DEBUG nova.compute.manager [req-fcaac219-1160-481d-ad2a-c967af33e601 req-03ce0079-6eb1-4783-b4b0-76bada62889a service nova] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Refreshing instance network info cache due to event network-changed-aae58430-6a65-4482-befa-f615cc7055d5. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 554.582983] env[61970]: DEBUG oslo_concurrency.lockutils [req-fcaac219-1160-481d-ad2a-c967af33e601 req-03ce0079-6eb1-4783-b4b0-76bada62889a service nova] Acquiring lock "refresh_cache-36c69772-6e47-4c00-a582-b46d233e13a1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.584410] env[61970]: DEBUG nova.network.neutron [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.724857] env[61970]: DEBUG nova.network.neutron [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.825343] env[61970]: DEBUG nova.network.neutron [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.068010] env[61970]: DEBUG nova.compute.manager [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 555.227795] env[61970]: INFO nova.compute.manager [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 10e1500f-70f1-40b6-aad6-5f9d75e1554b] Took 1.10 seconds to deallocate network for instance. [ 555.264803] env[61970]: DEBUG nova.network.neutron [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.284388] env[61970]: DEBUG nova.network.neutron [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.500843] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9720626a-64b0-46d0-9d16-d5970dcc1553 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.513825] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-918f4c16-f2b7-4946-982e-ca11a78b5f6a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.553524] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ac218e-a5ea-4d67-afc8-788ca8ff26eb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.566957] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd2c6ed-97f0-412c-8b5c-5d02302905b6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.595910] env[61970]: DEBUG nova.compute.provider_tree [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 555.599548] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.706014] env[61970]: DEBUG nova.compute.manager [None req-e36a0c09-f897-47b9-8bc5-cec7a81049cc tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 555.706957] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c8525f-ee2d-4ef7-afd0-202baad3ee63 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.769339] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Releasing lock "refresh_cache-36c69772-6e47-4c00-a582-b46d233e13a1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.769824] env[61970]: DEBUG nova.compute.manager [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 555.770620] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 555.774144] env[61970]: DEBUG oslo_concurrency.lockutils [req-fcaac219-1160-481d-ad2a-c967af33e601 req-03ce0079-6eb1-4783-b4b0-76bada62889a service nova] Acquired lock "refresh_cache-36c69772-6e47-4c00-a582-b46d233e13a1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.774144] env[61970]: DEBUG nova.network.neutron [req-fcaac219-1160-481d-ad2a-c967af33e601 req-03ce0079-6eb1-4783-b4b0-76bada62889a service nova] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Refreshing network info cache for port aae58430-6a65-4482-befa-f615cc7055d5 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 555.775295] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2b4ab4ed-a688-4f84-9303-1d5b85440671 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.789427] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f5ec8a-9774-4c1f-a9a4-bb51e666b796 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.809690] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Releasing lock "refresh_cache-0b34c674-6ef0-4c30-83e9-b4a5623ea044" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.809787] env[61970]: DEBUG nova.compute.manager [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 555.809919] env[61970]: DEBUG nova.compute.manager [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 555.810094] env[61970]: DEBUG nova.network.neutron [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 555.826589] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 36c69772-6e47-4c00-a582-b46d233e13a1 could not be found. [ 555.826851] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 555.827079] env[61970]: INFO nova.compute.manager [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Took 0.06 seconds to destroy the instance on the hypervisor. [ 555.827420] env[61970]: DEBUG oslo.service.loopingcall [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 555.827652] env[61970]: DEBUG nova.compute.manager [-] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 555.827780] env[61970]: DEBUG nova.network.neutron [-] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 555.907957] env[61970]: DEBUG nova.network.neutron [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.974480] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Acquiring lock "de9ed575-9e43-4c67-89ed-0549b43fb81a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.974779] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Lock "de9ed575-9e43-4c67-89ed-0549b43fb81a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.990100] env[61970]: DEBUG nova.network.neutron [-] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.104020] env[61970]: DEBUG nova.scheduler.client.report [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 556.222196] env[61970]: INFO nova.compute.manager [None req-e36a0c09-f897-47b9-8bc5-cec7a81049cc tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] instance snapshotting [ 556.222196] env[61970]: DEBUG nova.objects.instance [None req-e36a0c09-f897-47b9-8bc5-cec7a81049cc tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Lazy-loading 'flavor' on Instance uuid c4d3b145-84be-4d19-afc7-4c29fa193760 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 556.278968] env[61970]: INFO nova.scheduler.client.report [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Deleted allocations for instance 10e1500f-70f1-40b6-aad6-5f9d75e1554b [ 556.384633] env[61970]: DEBUG nova.network.neutron [req-fcaac219-1160-481d-ad2a-c967af33e601 req-03ce0079-6eb1-4783-b4b0-76bada62889a service nova] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.410062] env[61970]: DEBUG nova.network.neutron [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.491520] env[61970]: DEBUG nova.network.neutron [-] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.606519] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.608107] env[61970]: DEBUG nova.compute.manager [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 556.610744] env[61970]: DEBUG oslo_concurrency.lockutils [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.535s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.729801] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-676d5845-39ca-491b-8b43-396baecdd820 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.751387] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42714003-6b33-4cb6-99a1-fab1617d4861 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.793734] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8f58a444-b42c-49b3-bb47-f17f1016ee26 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "10e1500f-70f1-40b6-aad6-5f9d75e1554b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.007s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.913251] env[61970]: INFO nova.compute.manager [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] [instance: 0b34c674-6ef0-4c30-83e9-b4a5623ea044] Took 1.10 seconds to deallocate network for instance. [ 556.937038] env[61970]: DEBUG nova.network.neutron [req-fcaac219-1160-481d-ad2a-c967af33e601 req-03ce0079-6eb1-4783-b4b0-76bada62889a service nova] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.996718] env[61970]: INFO nova.compute.manager [-] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Took 1.17 seconds to deallocate network for instance. [ 557.003024] env[61970]: DEBUG nova.compute.claims [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 557.003358] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.117368] env[61970]: DEBUG nova.compute.utils [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 557.127151] env[61970]: DEBUG nova.compute.manager [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Not allocating networking since 'none' was specified. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 557.264909] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e36a0c09-f897-47b9-8bc5-cec7a81049cc tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Creating Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 557.269498] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2f92e722-a6f7-4b18-8471-aa98814e101f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.279863] env[61970]: DEBUG oslo_vmware.api [None req-e36a0c09-f897-47b9-8bc5-cec7a81049cc tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Waiting for the task: (returnval){ [ 557.279863] env[61970]: value = "task-1355472" [ 557.279863] env[61970]: _type = "Task" [ 557.279863] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.292034] env[61970]: DEBUG oslo_vmware.api [None req-e36a0c09-f897-47b9-8bc5-cec7a81049cc tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355472, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.298957] env[61970]: DEBUG nova.compute.manager [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 557.411111] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c3155c-caca-49a6-baeb-50172fef821d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.424060] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749998d3-be64-4b10-b5bd-8756b92eb838 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.459498] env[61970]: DEBUG oslo_concurrency.lockutils [req-fcaac219-1160-481d-ad2a-c967af33e601 req-03ce0079-6eb1-4783-b4b0-76bada62889a service nova] Releasing lock "refresh_cache-36c69772-6e47-4c00-a582-b46d233e13a1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.460722] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212b372c-d20d-4111-aa59-5bad3efb947c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.470393] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59831753-6b35-416e-9dee-f56566973304 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.485444] env[61970]: DEBUG nova.compute.provider_tree [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 557.538572] env[61970]: ERROR nova.compute.manager [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c09b9037-d99d-4c4e-950b-27276f22b702, please check neutron logs for more information. [ 557.538572] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 557.538572] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.538572] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 557.538572] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.538572] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 557.538572] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.538572] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 557.538572] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.538572] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 557.538572] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.538572] env[61970]: ERROR nova.compute.manager raise self.value [ 557.538572] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.538572] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 557.538572] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.538572] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 557.539289] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.539289] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 557.539289] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c09b9037-d99d-4c4e-950b-27276f22b702, please check neutron logs for more information. [ 557.539289] env[61970]: ERROR nova.compute.manager [ 557.539289] env[61970]: Traceback (most recent call last): [ 557.539289] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 557.539289] env[61970]: listener.cb(fileno) [ 557.539289] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.539289] env[61970]: result = function(*args, **kwargs) [ 557.539289] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 557.539289] env[61970]: return func(*args, **kwargs) [ 557.539289] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.539289] env[61970]: raise e [ 557.539289] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.539289] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 557.539289] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.539289] env[61970]: created_port_ids = self._update_ports_for_instance( [ 557.539289] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.539289] env[61970]: with excutils.save_and_reraise_exception(): [ 557.539289] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.539289] env[61970]: self.force_reraise() [ 557.539289] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.539289] env[61970]: raise self.value [ 557.539289] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.539289] env[61970]: updated_port = self._update_port( [ 557.539289] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.539289] env[61970]: _ensure_no_port_binding_failure(port) [ 557.539289] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.539289] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 557.541013] env[61970]: nova.exception.PortBindingFailed: Binding failed for port c09b9037-d99d-4c4e-950b-27276f22b702, please check neutron logs for more information. [ 557.541013] env[61970]: Removing descriptor: 15 [ 557.541013] env[61970]: ERROR nova.compute.manager [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c09b9037-d99d-4c4e-950b-27276f22b702, please check neutron logs for more information. [ 557.541013] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Traceback (most recent call last): [ 557.541013] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 557.541013] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] yield resources [ 557.541013] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 557.541013] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] self.driver.spawn(context, instance, image_meta, [ 557.541013] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 557.541013] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.541013] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.541013] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] vm_ref = self.build_virtual_machine(instance, [ 557.542662] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.542662] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.542662] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.542662] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] for vif in network_info: [ 557.542662] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 557.542662] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] return self._sync_wrapper(fn, *args, **kwargs) [ 557.542662] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 557.542662] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] self.wait() [ 557.542662] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 557.542662] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] self[:] = self._gt.wait() [ 557.542662] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.542662] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] return self._exit_event.wait() [ 557.542662] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.543356] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] result = hub.switch() [ 557.543356] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.543356] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] return self.greenlet.switch() [ 557.543356] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.543356] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] result = function(*args, **kwargs) [ 557.543356] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 557.543356] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] return func(*args, **kwargs) [ 557.543356] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.543356] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] raise e [ 557.543356] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.543356] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] nwinfo = self.network_api.allocate_for_instance( [ 557.543356] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.543356] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] created_port_ids = self._update_ports_for_instance( [ 557.543937] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.543937] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] with excutils.save_and_reraise_exception(): [ 557.543937] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.543937] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] self.force_reraise() [ 557.543937] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.543937] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] raise self.value [ 557.543937] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.543937] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] updated_port = self._update_port( [ 557.543937] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.543937] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] _ensure_no_port_binding_failure(port) [ 557.543937] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.543937] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] raise exception.PortBindingFailed(port_id=port['id']) [ 557.544626] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] nova.exception.PortBindingFailed: Binding failed for port c09b9037-d99d-4c4e-950b-27276f22b702, please check neutron logs for more information. [ 557.544626] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] [ 557.544626] env[61970]: INFO nova.compute.manager [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Terminating instance [ 557.544936] env[61970]: DEBUG oslo_concurrency.lockutils [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Acquiring lock "refresh_cache-e3744562-c9ab-4902-880c-6f0153b8a183" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.545059] env[61970]: DEBUG oslo_concurrency.lockutils [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Acquired lock "refresh_cache-e3744562-c9ab-4902-880c-6f0153b8a183" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.545258] env[61970]: DEBUG nova.network.neutron [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 557.626882] env[61970]: DEBUG nova.compute.manager [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 557.716042] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Acquiring lock "c4d3b145-84be-4d19-afc7-4c29fa193760" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.716343] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Lock "c4d3b145-84be-4d19-afc7-4c29fa193760" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.717056] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Acquiring lock "c4d3b145-84be-4d19-afc7-4c29fa193760-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.717056] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Lock "c4d3b145-84be-4d19-afc7-4c29fa193760-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.717185] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Lock "c4d3b145-84be-4d19-afc7-4c29fa193760-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.720502] env[61970]: INFO nova.compute.manager [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Terminating instance [ 557.725480] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Acquiring lock "refresh_cache-c4d3b145-84be-4d19-afc7-4c29fa193760" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.725631] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Acquired lock "refresh_cache-c4d3b145-84be-4d19-afc7-4c29fa193760" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.725793] env[61970]: DEBUG nova.network.neutron [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 557.789645] env[61970]: DEBUG oslo_vmware.api [None req-e36a0c09-f897-47b9-8bc5-cec7a81049cc tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355472, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.838500] env[61970]: DEBUG oslo_concurrency.lockutils [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.989347] env[61970]: DEBUG nova.scheduler.client.report [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 557.993706] env[61970]: INFO nova.scheduler.client.report [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Deleted allocations for instance 0b34c674-6ef0-4c30-83e9-b4a5623ea044 [ 558.117590] env[61970]: DEBUG nova.network.neutron [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.296596] env[61970]: DEBUG oslo_vmware.api [None req-e36a0c09-f897-47b9-8bc5-cec7a81049cc tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355472, 'name': CreateSnapshot_Task, 'duration_secs': 0.908254} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.297038] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e36a0c09-f897-47b9-8bc5-cec7a81049cc tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Created Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 558.299973] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3184139-0eeb-40f1-b91f-5351e0c06f77 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.303798] env[61970]: DEBUG nova.network.neutron [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.391147] env[61970]: DEBUG nova.compute.manager [req-c85c07ba-225e-4b33-ba75-15348063fb54 req-3c419ef4-c413-43cf-a0e1-bda384210aa0 service nova] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Received event network-vif-deleted-aae58430-6a65-4482-befa-f615cc7055d5 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 558.391147] env[61970]: DEBUG nova.compute.manager [req-c85c07ba-225e-4b33-ba75-15348063fb54 req-3c419ef4-c413-43cf-a0e1-bda384210aa0 service nova] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Received event network-changed-c09b9037-d99d-4c4e-950b-27276f22b702 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 558.391312] env[61970]: DEBUG nova.compute.manager [req-c85c07ba-225e-4b33-ba75-15348063fb54 req-3c419ef4-c413-43cf-a0e1-bda384210aa0 service nova] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Refreshing instance network info cache due to event network-changed-c09b9037-d99d-4c4e-950b-27276f22b702. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 558.391440] env[61970]: DEBUG oslo_concurrency.lockutils [req-c85c07ba-225e-4b33-ba75-15348063fb54 req-3c419ef4-c413-43cf-a0e1-bda384210aa0 service nova] Acquiring lock "refresh_cache-e3744562-c9ab-4902-880c-6f0153b8a183" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.447191] env[61970]: DEBUG nova.network.neutron [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.475348] env[61970]: DEBUG nova.network.neutron [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.505295] env[61970]: DEBUG oslo_concurrency.lockutils [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.894s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.505915] env[61970]: ERROR nova.compute.manager [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e544516b-8f51-45de-9462-e55eae71c56b, please check neutron logs for more information. [ 558.505915] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Traceback (most recent call last): [ 558.505915] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 558.505915] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] self.driver.spawn(context, instance, image_meta, [ 558.505915] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 558.505915] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 558.505915] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 558.505915] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] vm_ref = self.build_virtual_machine(instance, [ 558.505915] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 558.505915] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] vif_infos = vmwarevif.get_vif_info(self._session, [ 558.505915] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 558.506276] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] for vif in network_info: [ 558.506276] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 558.506276] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] return self._sync_wrapper(fn, *args, **kwargs) [ 558.506276] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 558.506276] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] self.wait() [ 558.506276] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 558.506276] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] self[:] = self._gt.wait() [ 558.506276] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 558.506276] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] return self._exit_event.wait() [ 558.506276] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 558.506276] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] result = hub.switch() [ 558.506276] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 558.506276] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] return self.greenlet.switch() [ 558.506699] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.506699] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] result = function(*args, **kwargs) [ 558.506699] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 558.506699] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] return func(*args, **kwargs) [ 558.506699] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 558.506699] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] raise e [ 558.506699] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.506699] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] nwinfo = self.network_api.allocate_for_instance( [ 558.506699] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.506699] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] created_port_ids = self._update_ports_for_instance( [ 558.506699] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.506699] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] with excutils.save_and_reraise_exception(): [ 558.506699] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.507094] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] self.force_reraise() [ 558.507094] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.507094] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] raise self.value [ 558.507094] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.507094] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] updated_port = self._update_port( [ 558.507094] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.507094] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] _ensure_no_port_binding_failure(port) [ 558.507094] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.507094] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] raise exception.PortBindingFailed(port_id=port['id']) [ 558.507094] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] nova.exception.PortBindingFailed: Binding failed for port e544516b-8f51-45de-9462-e55eae71c56b, please check neutron logs for more information. [ 558.507094] env[61970]: ERROR nova.compute.manager [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] [ 558.507372] env[61970]: DEBUG nova.compute.utils [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Binding failed for port e544516b-8f51-45de-9462-e55eae71c56b, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 558.508582] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebc1963c-dc84-45f4-ab92-03e3856b2b33 tempest-ServersAdminNegativeTestJSON-1341636256 tempest-ServersAdminNegativeTestJSON-1341636256-project-member] Lock "0b34c674-6ef0-4c30-83e9-b4a5623ea044" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.541s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.508994] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.224s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.510671] env[61970]: INFO nova.compute.claims [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 558.518152] env[61970]: DEBUG nova.compute.manager [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Build of instance 4d9ee781-bca8-4cad-9fc5-87cb785d478e was re-scheduled: Binding failed for port e544516b-8f51-45de-9462-e55eae71c56b, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 558.520598] env[61970]: DEBUG nova.compute.manager [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 558.520845] env[61970]: DEBUG oslo_concurrency.lockutils [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Acquiring lock "refresh_cache-4d9ee781-bca8-4cad-9fc5-87cb785d478e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.520992] env[61970]: DEBUG oslo_concurrency.lockutils [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Acquired lock "refresh_cache-4d9ee781-bca8-4cad-9fc5-87cb785d478e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.521170] env[61970]: DEBUG nova.network.neutron [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 558.592209] env[61970]: DEBUG oslo_concurrency.lockutils [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Acquiring lock "0ef75e57-83af-4440-890e-95dcdb9920d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.592521] env[61970]: DEBUG oslo_concurrency.lockutils [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Lock "0ef75e57-83af-4440-890e-95dcdb9920d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.640948] env[61970]: DEBUG nova.compute.manager [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 558.689013] env[61970]: DEBUG nova.virt.hardware [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 558.689292] env[61970]: DEBUG nova.virt.hardware [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 558.689534] env[61970]: DEBUG nova.virt.hardware [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 558.689634] env[61970]: DEBUG nova.virt.hardware [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 558.689777] env[61970]: DEBUG nova.virt.hardware [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 558.689921] env[61970]: DEBUG nova.virt.hardware [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 558.690142] env[61970]: DEBUG nova.virt.hardware [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 558.690305] env[61970]: DEBUG nova.virt.hardware [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 558.694188] env[61970]: DEBUG nova.virt.hardware [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 558.694410] env[61970]: DEBUG nova.virt.hardware [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 558.694591] env[61970]: DEBUG nova.virt.hardware [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 558.695811] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd5a4568-4ebe-4483-b0a6-4c03be73ab89 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.703211] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc8aa91-f55b-466e-96bb-4d389b09017c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.717555] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Instance VIF info [] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 558.722854] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Creating folder: Project (d9104d6d950a44ed88064533d8efb88e). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 558.723639] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bd04a3b3-61f8-4483-9f13-1ceb7339fa79 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.733472] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Created folder: Project (d9104d6d950a44ed88064533d8efb88e) in parent group-v288740. [ 558.733652] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Creating folder: Instances. Parent ref: group-v288745. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 558.735055] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c66a2415-6b1f-469b-95d9-790d6c6fe47e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.742477] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Created folder: Instances in parent group-v288745. [ 558.742784] env[61970]: DEBUG oslo.service.loopingcall [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 558.742985] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 558.743110] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f0f1c49d-4b01-468a-abe6-e862cd8591e1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.759955] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 558.759955] env[61970]: value = "task-1355475" [ 558.759955] env[61970]: _type = "Task" [ 558.759955] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.767623] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355475, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.812518] env[61970]: DEBUG nova.compute.manager [None req-e36a0c09-f897-47b9-8bc5-cec7a81049cc tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Instance disappeared during snapshot {{(pid=61970) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 558.951772] env[61970]: DEBUG oslo_concurrency.lockutils [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Releasing lock "refresh_cache-e3744562-c9ab-4902-880c-6f0153b8a183" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.951772] env[61970]: DEBUG nova.compute.manager [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 558.951772] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 558.951910] env[61970]: DEBUG oslo_concurrency.lockutils [req-c85c07ba-225e-4b33-ba75-15348063fb54 req-3c419ef4-c413-43cf-a0e1-bda384210aa0 service nova] Acquired lock "refresh_cache-e3744562-c9ab-4902-880c-6f0153b8a183" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.952043] env[61970]: DEBUG nova.network.neutron [req-c85c07ba-225e-4b33-ba75-15348063fb54 req-3c419ef4-c413-43cf-a0e1-bda384210aa0 service nova] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Refreshing network info cache for port c09b9037-d99d-4c4e-950b-27276f22b702 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 558.953166] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2b9bae34-a8b6-4d16-872e-9ad5408883c6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.966666] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f26454f6-dad1-4b5a-82e2-319c0ff1bceb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.980938] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Releasing lock "refresh_cache-c4d3b145-84be-4d19-afc7-4c29fa193760" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.981408] env[61970]: DEBUG nova.compute.manager [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 558.981616] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 558.982681] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c400289-617d-4d28-bb0b-2bb25e25f621 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.990927] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 558.991210] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-81ed1ac2-f0e2-4297-a446-7273e40b0468 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.001658] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e3744562-c9ab-4902-880c-6f0153b8a183 could not be found. [ 559.001658] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 559.001658] env[61970]: INFO nova.compute.manager [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Took 0.05 seconds to destroy the instance on the hypervisor. [ 559.001837] env[61970]: DEBUG oslo.service.loopingcall [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 559.003746] env[61970]: DEBUG nova.compute.manager [-] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.003850] env[61970]: DEBUG nova.network.neutron [-] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 559.005773] env[61970]: DEBUG oslo_vmware.api [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Waiting for the task: (returnval){ [ 559.005773] env[61970]: value = "task-1355476" [ 559.005773] env[61970]: _type = "Task" [ 559.005773] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.010557] env[61970]: DEBUG nova.compute.manager [None req-e36a0c09-f897-47b9-8bc5-cec7a81049cc tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Found 0 images (rotation: 2) {{(pid=61970) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 559.018733] env[61970]: DEBUG oslo_vmware.api [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355476, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.029862] env[61970]: DEBUG nova.compute.manager [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 559.038874] env[61970]: DEBUG nova.network.neutron [-] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.073024] env[61970]: DEBUG nova.network.neutron [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.170232] env[61970]: DEBUG nova.network.neutron [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.274993] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355475, 'name': CreateVM_Task, 'duration_secs': 0.411663} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.274993] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 559.278272] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.278272] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.278272] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 559.278272] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc968204-9228-4adc-a634-02030f524c4c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.287688] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Waiting for the task: (returnval){ [ 559.287688] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e48b74-8563-2100-7e59-796d602309c1" [ 559.287688] env[61970]: _type = "Task" [ 559.287688] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.297026] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e48b74-8563-2100-7e59-796d602309c1, 'name': SearchDatastore_Task, 'duration_secs': 0.008655} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.297026] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.297361] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 559.298180] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.298418] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.298661] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 559.298949] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5effa16e-bacb-450a-a62c-5f9661a93a54 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.308346] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 559.308533] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 559.309276] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ec9585c-b5ed-4a83-93ec-a1e463e2a10f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.314766] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Waiting for the task: (returnval){ [ 559.314766] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5264ba59-f8a6-fb60-a289-e49ffb730ac6" [ 559.314766] env[61970]: _type = "Task" [ 559.314766] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.325107] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5264ba59-f8a6-fb60-a289-e49ffb730ac6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.499626] env[61970]: DEBUG nova.network.neutron [req-c85c07ba-225e-4b33-ba75-15348063fb54 req-3c419ef4-c413-43cf-a0e1-bda384210aa0 service nova] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.515504] env[61970]: DEBUG oslo_vmware.api [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355476, 'name': PowerOffVM_Task, 'duration_secs': 0.190188} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.515807] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 559.516073] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 559.516268] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-79e0f25d-5f6a-44fe-9e79-d64f37cd0074 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.539858] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 559.540187] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 559.540781] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Deleting the datastore file [datastore2] c4d3b145-84be-4d19-afc7-4c29fa193760 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 559.543101] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14cd055f-d72d-4220-89f5-fb2aea85d09d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.544896] env[61970]: DEBUG nova.network.neutron [-] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.551372] env[61970]: DEBUG oslo_vmware.api [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Waiting for the task: (returnval){ [ 559.551372] env[61970]: value = "task-1355478" [ 559.551372] env[61970]: _type = "Task" [ 559.551372] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.560952] env[61970]: DEBUG oslo_vmware.api [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355478, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.576803] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.676024] env[61970]: DEBUG oslo_concurrency.lockutils [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Releasing lock "refresh_cache-4d9ee781-bca8-4cad-9fc5-87cb785d478e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.676024] env[61970]: DEBUG nova.compute.manager [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 559.676024] env[61970]: DEBUG nova.compute.manager [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.676024] env[61970]: DEBUG nova.network.neutron [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 559.676024] env[61970]: DEBUG nova.network.neutron [req-c85c07ba-225e-4b33-ba75-15348063fb54 req-3c419ef4-c413-43cf-a0e1-bda384210aa0 service nova] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.701133] env[61970]: DEBUG nova.network.neutron [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.770916] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06580d81-809f-4850-b00e-125dddde7b04 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.778938] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751d6f82-bfb4-4c3e-b3b1-a687e3a6d9c2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.813899] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e038a691-64ee-4217-b8ce-f89b83dfea0e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.827008] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba7deb4-408c-4624-b75b-804affd3ffb3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.831915] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5264ba59-f8a6-fb60-a289-e49ffb730ac6, 'name': SearchDatastore_Task, 'duration_secs': 0.007546} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.833569] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8b41234-1798-43f8-8f08-da9b4de3d268 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.845592] env[61970]: DEBUG nova.compute.provider_tree [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 559.849352] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Waiting for the task: (returnval){ [ 559.849352] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52fa3905-9ba1-3ef5-ebcb-054092b91e47" [ 559.849352] env[61970]: _type = "Task" [ 559.849352] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.858212] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52fa3905-9ba1-3ef5-ebcb-054092b91e47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.050251] env[61970]: INFO nova.compute.manager [-] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Took 1.04 seconds to deallocate network for instance. [ 560.051505] env[61970]: DEBUG nova.compute.claims [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 560.051715] env[61970]: DEBUG oslo_concurrency.lockutils [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.065753] env[61970]: DEBUG oslo_vmware.api [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Task: {'id': task-1355478, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14021} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.067670] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 560.067670] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 560.067670] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 560.067670] env[61970]: INFO nova.compute.manager [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Took 1.09 seconds to destroy the instance on the hypervisor. [ 560.067670] env[61970]: DEBUG oslo.service.loopingcall [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 560.067949] env[61970]: DEBUG nova.compute.manager [-] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 560.067949] env[61970]: DEBUG nova.network.neutron [-] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 560.125282] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Acquiring lock "d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.125779] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Lock "d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.136713] env[61970]: DEBUG nova.network.neutron [-] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.182269] env[61970]: DEBUG oslo_concurrency.lockutils [req-c85c07ba-225e-4b33-ba75-15348063fb54 req-3c419ef4-c413-43cf-a0e1-bda384210aa0 service nova] Releasing lock "refresh_cache-e3744562-c9ab-4902-880c-6f0153b8a183" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.203356] env[61970]: DEBUG nova.network.neutron [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.353630] env[61970]: DEBUG nova.scheduler.client.report [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 560.371329] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52fa3905-9ba1-3ef5-ebcb-054092b91e47, 'name': SearchDatastore_Task, 'duration_secs': 0.009307} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.371886] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.372335] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 8881014c-8505-4e75-8c31-e25b6a8d2cd4/8881014c-8505-4e75-8c31-e25b6a8d2cd4.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 560.373737] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-75dd69d3-ad0a-4ab3-ba35-ef84f8ffe02b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.387606] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Waiting for the task: (returnval){ [ 560.387606] env[61970]: value = "task-1355479" [ 560.387606] env[61970]: _type = "Task" [ 560.387606] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.399293] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': task-1355479, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.641772] env[61970]: DEBUG nova.network.neutron [-] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.708423] env[61970]: INFO nova.compute.manager [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] [instance: 4d9ee781-bca8-4cad-9fc5-87cb785d478e] Took 1.03 seconds to deallocate network for instance. [ 560.861449] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.353s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.861603] env[61970]: DEBUG nova.compute.manager [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 560.865677] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.315s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.901577] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': task-1355479, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.144159] env[61970]: INFO nova.compute.manager [-] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Took 1.08 seconds to deallocate network for instance. [ 561.371752] env[61970]: DEBUG nova.compute.utils [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 561.382044] env[61970]: DEBUG nova.compute.manager [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 561.382044] env[61970]: DEBUG nova.network.neutron [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 561.401895] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': task-1355479, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.537871} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.402532] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 8881014c-8505-4e75-8c31-e25b6a8d2cd4/8881014c-8505-4e75-8c31-e25b6a8d2cd4.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 561.402889] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 561.403628] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d8760ae4-438c-47c4-8035-49cb63df252a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.412835] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Waiting for the task: (returnval){ [ 561.412835] env[61970]: value = "task-1355484" [ 561.412835] env[61970]: _type = "Task" [ 561.412835] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.422543] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': task-1355484, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.488408] env[61970]: DEBUG oslo_concurrency.lockutils [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Acquiring lock "44ea1800-005c-485f-969d-4bf71d58c284" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.488643] env[61970]: DEBUG oslo_concurrency.lockutils [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Lock "44ea1800-005c-485f-969d-4bf71d58c284" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.494311] env[61970]: DEBUG nova.policy [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '627f87a9d8754a95a6a99cc02731bf2a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e0d60daa756447dbe58c5b1c5171a71', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 561.656374] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.761298] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Acquiring lock "59d2c196-9230-4303-9b99-92211c53cb15" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.761535] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Lock "59d2c196-9230-4303-9b99-92211c53cb15" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.785429] env[61970]: INFO nova.scheduler.client.report [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Deleted allocations for instance 4d9ee781-bca8-4cad-9fc5-87cb785d478e [ 561.881896] env[61970]: DEBUG nova.compute.manager [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 561.920459] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance c4d3b145-84be-4d19-afc7-4c29fa193760 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 561.920827] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance bce7d001-2885-409c-b455-4c8a3ce8eb9e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 561.921057] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 36c69772-6e47-4c00-a582-b46d233e13a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 561.921421] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance e3744562-c9ab-4902-880c-6f0153b8a183 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 561.922695] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 8881014c-8505-4e75-8c31-e25b6a8d2cd4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 561.923056] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 561.936727] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': task-1355484, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07569} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.937020] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 561.937914] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8380dc28-6365-457f-ae0b-56e8c9f6e709 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.962781] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Reconfiguring VM instance instance-00000009 to attach disk [datastore2] 8881014c-8505-4e75-8c31-e25b6a8d2cd4/8881014c-8505-4e75-8c31-e25b6a8d2cd4.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 561.964147] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b926853-82c8-4494-87a0-08b32732a2ab {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.989321] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Waiting for the task: (returnval){ [ 561.989321] env[61970]: value = "task-1355485" [ 561.989321] env[61970]: _type = "Task" [ 561.989321] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.998954] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': task-1355485, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.293784] env[61970]: DEBUG oslo_concurrency.lockutils [None req-844a0404-34d4-4e67-b5e4-db5d0a43db20 tempest-ServerExternalEventsTest-808194474 tempest-ServerExternalEventsTest-808194474-project-member] Lock "4d9ee781-bca8-4cad-9fc5-87cb785d478e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.600s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.412934] env[61970]: DEBUG nova.network.neutron [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Successfully created port: 9886c8df-135a-4a75-bac1-b546e16ef653 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 562.427095] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 39bc8558-ce7f-41e9-b913-9b628a220ca7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 562.500503] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': task-1355485, 'name': ReconfigVM_Task, 'duration_secs': 0.277126} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.500944] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Reconfigured VM instance instance-00000009 to attach disk [datastore2] 8881014c-8505-4e75-8c31-e25b6a8d2cd4/8881014c-8505-4e75-8c31-e25b6a8d2cd4.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 562.501666] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ec016389-2382-4197-a33f-54933062730b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.508506] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Waiting for the task: (returnval){ [ 562.508506] env[61970]: value = "task-1355486" [ 562.508506] env[61970]: _type = "Task" [ 562.508506] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.517940] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': task-1355486, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.589760] env[61970]: DEBUG nova.compute.manager [req-4765deb4-48ea-4e22-84ab-661d8d0753b3 req-df9f464d-494a-4597-8d23-22ed1783f94c service nova] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Received event network-vif-deleted-c09b9037-d99d-4c4e-950b-27276f22b702 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 562.796678] env[61970]: DEBUG nova.compute.manager [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 562.894472] env[61970]: DEBUG nova.compute.manager [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 562.929186] env[61970]: DEBUG nova.virt.hardware [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:38:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='658270470',id=21,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-827298559',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 562.930088] env[61970]: DEBUG nova.virt.hardware [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 562.930088] env[61970]: DEBUG nova.virt.hardware [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 562.930181] env[61970]: DEBUG nova.virt.hardware [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 562.930329] env[61970]: DEBUG nova.virt.hardware [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 562.930442] env[61970]: DEBUG nova.virt.hardware [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 562.930654] env[61970]: DEBUG nova.virt.hardware [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 562.930810] env[61970]: DEBUG nova.virt.hardware [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 562.930976] env[61970]: DEBUG nova.virt.hardware [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 562.931150] env[61970]: DEBUG nova.virt.hardware [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 562.931318] env[61970]: DEBUG nova.virt.hardware [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 562.933792] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 17f41a87-4fd0-4866-905a-2aa3d2e9691b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 562.933871] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8122871-1175-4247-8fba-e55d192335da {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.943348] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69943eb4-fb8f-4bb5-a3d3-10a34ef7d9a3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.019955] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': task-1355486, 'name': Rename_Task, 'duration_secs': 0.129627} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 563.024412] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 563.025077] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-004fca74-4cb8-4971-a486-6b68f4bf2a4f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.033268] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Waiting for the task: (returnval){ [ 563.033268] env[61970]: value = "task-1355487" [ 563.033268] env[61970]: _type = "Task" [ 563.033268] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.047604] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': task-1355487, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.344934] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.439555] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 8bacc19d-2626-406a-a3fb-f1069674963c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 563.453636] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "7c48e5b6-eb5f-4c35-9966-1cfc69182595" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.454023] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "7c48e5b6-eb5f-4c35-9966-1cfc69182595" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.550144] env[61970]: DEBUG oslo_vmware.api [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': task-1355487, 'name': PowerOnVM_Task, 'duration_secs': 0.459858} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 563.550459] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 563.550666] env[61970]: INFO nova.compute.manager [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Took 4.91 seconds to spawn the instance on the hypervisor. [ 563.550845] env[61970]: DEBUG nova.compute.manager [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 563.553349] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-658dcc4c-91fe-4f81-a7fe-4cef3dba42d7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.754031] env[61970]: ERROR nova.compute.manager [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9886c8df-135a-4a75-bac1-b546e16ef653, please check neutron logs for more information. [ 563.754031] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 563.754031] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.754031] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 563.754031] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.754031] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 563.754031] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.754031] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 563.754031] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.754031] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 563.754031] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.754031] env[61970]: ERROR nova.compute.manager raise self.value [ 563.754031] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.754031] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 563.754031] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.754031] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 563.754688] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.754688] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 563.754688] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9886c8df-135a-4a75-bac1-b546e16ef653, please check neutron logs for more information. [ 563.754688] env[61970]: ERROR nova.compute.manager [ 563.754688] env[61970]: Traceback (most recent call last): [ 563.754688] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 563.754688] env[61970]: listener.cb(fileno) [ 563.754688] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.754688] env[61970]: result = function(*args, **kwargs) [ 563.754688] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 563.754688] env[61970]: return func(*args, **kwargs) [ 563.754688] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.754688] env[61970]: raise e [ 563.754688] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.754688] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 563.754688] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.754688] env[61970]: created_port_ids = self._update_ports_for_instance( [ 563.754688] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.754688] env[61970]: with excutils.save_and_reraise_exception(): [ 563.754688] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.754688] env[61970]: self.force_reraise() [ 563.754688] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.754688] env[61970]: raise self.value [ 563.754688] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.754688] env[61970]: updated_port = self._update_port( [ 563.754688] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.754688] env[61970]: _ensure_no_port_binding_failure(port) [ 563.754688] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.754688] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 563.755479] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 9886c8df-135a-4a75-bac1-b546e16ef653, please check neutron logs for more information. [ 563.755479] env[61970]: Removing descriptor: 15 [ 563.755479] env[61970]: ERROR nova.compute.manager [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9886c8df-135a-4a75-bac1-b546e16ef653, please check neutron logs for more information. [ 563.755479] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Traceback (most recent call last): [ 563.755479] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 563.755479] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] yield resources [ 563.755479] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 563.755479] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] self.driver.spawn(context, instance, image_meta, [ 563.755479] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 563.755479] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 563.755479] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 563.755479] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] vm_ref = self.build_virtual_machine(instance, [ 563.755888] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 563.755888] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] vif_infos = vmwarevif.get_vif_info(self._session, [ 563.755888] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 563.755888] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] for vif in network_info: [ 563.755888] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 563.755888] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] return self._sync_wrapper(fn, *args, **kwargs) [ 563.755888] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 563.755888] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] self.wait() [ 563.755888] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 563.755888] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] self[:] = self._gt.wait() [ 563.755888] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 563.755888] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] return self._exit_event.wait() [ 563.755888] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 563.756351] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] result = hub.switch() [ 563.756351] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 563.756351] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] return self.greenlet.switch() [ 563.756351] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.756351] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] result = function(*args, **kwargs) [ 563.756351] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 563.756351] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] return func(*args, **kwargs) [ 563.756351] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.756351] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] raise e [ 563.756351] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.756351] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] nwinfo = self.network_api.allocate_for_instance( [ 563.756351] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.756351] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] created_port_ids = self._update_ports_for_instance( [ 563.756804] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.756804] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] with excutils.save_and_reraise_exception(): [ 563.756804] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.756804] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] self.force_reraise() [ 563.756804] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.756804] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] raise self.value [ 563.756804] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.756804] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] updated_port = self._update_port( [ 563.756804] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.756804] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] _ensure_no_port_binding_failure(port) [ 563.756804] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.756804] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] raise exception.PortBindingFailed(port_id=port['id']) [ 563.757189] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] nova.exception.PortBindingFailed: Binding failed for port 9886c8df-135a-4a75-bac1-b546e16ef653, please check neutron logs for more information. [ 563.757189] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] [ 563.757189] env[61970]: INFO nova.compute.manager [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Terminating instance [ 563.764700] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Acquiring lock "refresh_cache-604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.764700] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Acquired lock "refresh_cache-604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.765535] env[61970]: DEBUG nova.network.neutron [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 563.773008] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Acquiring lock "a6a3b8fe-6911-4d8f-bd3e-8044c47a541b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.773255] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Lock "a6a3b8fe-6911-4d8f-bd3e-8044c47a541b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.949728] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance fb0aec23-e010-4f9c-8a1d-1b31d17190b8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 564.079084] env[61970]: INFO nova.compute.manager [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Took 19.12 seconds to build instance. [ 564.321209] env[61970]: DEBUG nova.network.neutron [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.455695] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance de9ed575-9e43-4c67-89ed-0549b43fb81a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 564.511531] env[61970]: DEBUG nova.network.neutron [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.581136] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1b029567-f16b-4281-b4d2-e6d3358be2c0 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Lock "8881014c-8505-4e75-8c31-e25b6a8d2cd4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.645s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.961165] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 0ef75e57-83af-4440-890e-95dcdb9920d2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 565.016474] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Releasing lock "refresh_cache-604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.016835] env[61970]: DEBUG nova.compute.manager [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 565.017058] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 565.017383] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7a03c64d-f7e4-46fc-b54c-d11fe45979ef {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.028543] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77855c0b-84e5-49f4-adf9-fcb78c626aef {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.059431] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f could not be found. [ 565.059719] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 565.059924] env[61970]: INFO nova.compute.manager [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 565.060231] env[61970]: DEBUG oslo.service.loopingcall [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 565.060535] env[61970]: DEBUG nova.compute.manager [-] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 565.060599] env[61970]: DEBUG nova.network.neutron [-] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 565.083788] env[61970]: DEBUG nova.compute.manager [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 565.099746] env[61970]: DEBUG nova.network.neutron [-] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.465689] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 565.602938] env[61970]: DEBUG nova.network.neutron [-] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.611718] env[61970]: DEBUG oslo_concurrency.lockutils [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.789359] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Acquiring lock "65dba389-321c-478d-9555-bcd9c318f6d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.789450] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Lock "65dba389-321c-478d-9555-bcd9c318f6d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.971034] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 44ea1800-005c-485f-969d-4bf71d58c284 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 566.107671] env[61970]: INFO nova.compute.manager [-] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Took 1.05 seconds to deallocate network for instance. [ 566.111533] env[61970]: DEBUG nova.compute.claims [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 566.111721] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.122019] env[61970]: DEBUG nova.compute.manager [None req-41fc046c-ed38-4d62-bf16-c0092d52ea70 tempest-ServerDiagnosticsV248Test-831209024 tempest-ServerDiagnosticsV248Test-831209024-project-admin] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 566.123296] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e75dc5-c001-4b0d-96d9-c61638de957a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.130748] env[61970]: INFO nova.compute.manager [None req-41fc046c-ed38-4d62-bf16-c0092d52ea70 tempest-ServerDiagnosticsV248Test-831209024 tempest-ServerDiagnosticsV248Test-831209024-project-admin] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Retrieving diagnostics [ 566.131522] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aa1c640-e013-4076-898f-29afb0bfc8e7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.297080] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Acquiring lock "0c45c5c0-f8db-4631-ba9a-271d8f9ddff4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.297080] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Lock "0c45c5c0-f8db-4631-ba9a-271d8f9ddff4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.301067] env[61970]: DEBUG nova.compute.manager [req-4108095b-c76b-427c-a248-baf69287baab req-50f44ea6-fd10-48a5-906d-5d5b9f4cf4fd service nova] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Received event network-changed-9886c8df-135a-4a75-bac1-b546e16ef653 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 566.301067] env[61970]: DEBUG nova.compute.manager [req-4108095b-c76b-427c-a248-baf69287baab req-50f44ea6-fd10-48a5-906d-5d5b9f4cf4fd service nova] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Refreshing instance network info cache due to event network-changed-9886c8df-135a-4a75-bac1-b546e16ef653. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 566.301067] env[61970]: DEBUG oslo_concurrency.lockutils [req-4108095b-c76b-427c-a248-baf69287baab req-50f44ea6-fd10-48a5-906d-5d5b9f4cf4fd service nova] Acquiring lock "refresh_cache-604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.301067] env[61970]: DEBUG oslo_concurrency.lockutils [req-4108095b-c76b-427c-a248-baf69287baab req-50f44ea6-fd10-48a5-906d-5d5b9f4cf4fd service nova] Acquired lock "refresh_cache-604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.301067] env[61970]: DEBUG nova.network.neutron [req-4108095b-c76b-427c-a248-baf69287baab req-50f44ea6-fd10-48a5-906d-5d5b9f4cf4fd service nova] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Refreshing network info cache for port 9886c8df-135a-4a75-bac1-b546e16ef653 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 566.473318] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 59d2c196-9230-4303-9b99-92211c53cb15 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 566.473583] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 566.473723] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 566.750274] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16470ac0-ad79-42be-bac9-4d3d3702881d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.757931] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-384022af-6af5-4f14-b383-a3217335caf0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.786913] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c9a2763-abbd-485b-a9f1-80720741a5c1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.794767] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aae526d-0a2b-4a43-b20e-7702ca11945e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.822276] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 566.829743] env[61970]: DEBUG nova.network.neutron [req-4108095b-c76b-427c-a248-baf69287baab req-50f44ea6-fd10-48a5-906d-5d5b9f4cf4fd service nova] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.938451] env[61970]: DEBUG nova.network.neutron [req-4108095b-c76b-427c-a248-baf69287baab req-50f44ea6-fd10-48a5-906d-5d5b9f4cf4fd service nova] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.326515] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 567.443602] env[61970]: DEBUG oslo_concurrency.lockutils [req-4108095b-c76b-427c-a248-baf69287baab req-50f44ea6-fd10-48a5-906d-5d5b9f4cf4fd service nova] Releasing lock "refresh_cache-604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.443824] env[61970]: DEBUG nova.compute.manager [req-4108095b-c76b-427c-a248-baf69287baab req-50f44ea6-fd10-48a5-906d-5d5b9f4cf4fd service nova] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Received event network-vif-deleted-9886c8df-135a-4a75-bac1-b546e16ef653 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 567.831258] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 567.831529] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.966s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.831766] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.425s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.957894] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Acquiring lock "816de9f9-a8a9-415a-a6c4-57baea86df51" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.958417] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Lock "816de9f9-a8a9-415a-a6c4-57baea86df51" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.684746] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7c433d-8f99-4fce-a38d-1e5913e45490 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.695528] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8448633d-dfe8-4df3-a551-744da51d1e1f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.728040] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29569d8c-3cc7-45c0-bcb3-f58b7177f3fa {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.735505] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f6ee112-9e53-40e7-b99f-9718a1f6c6b6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.750727] env[61970]: DEBUG nova.compute.provider_tree [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.259086] env[61970]: DEBUG nova.scheduler.client.report [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 569.774809] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.940s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.774809] env[61970]: ERROR nova.compute.manager [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3b92f861-eeb4-4364-9ab4-21f10dce1713, please check neutron logs for more information. [ 569.774809] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Traceback (most recent call last): [ 569.774809] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 569.774809] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] self.driver.spawn(context, instance, image_meta, [ 569.774809] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 569.774809] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 569.774809] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 569.774809] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] vm_ref = self.build_virtual_machine(instance, [ 569.775465] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 569.775465] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] vif_infos = vmwarevif.get_vif_info(self._session, [ 569.775465] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 569.775465] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] for vif in network_info: [ 569.775465] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 569.775465] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] return self._sync_wrapper(fn, *args, **kwargs) [ 569.775465] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 569.775465] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] self.wait() [ 569.775465] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 569.775465] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] self[:] = self._gt.wait() [ 569.775465] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 569.775465] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] return self._exit_event.wait() [ 569.775465] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 569.775911] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] result = hub.switch() [ 569.775911] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 569.775911] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] return self.greenlet.switch() [ 569.775911] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.775911] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] result = function(*args, **kwargs) [ 569.775911] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 569.775911] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] return func(*args, **kwargs) [ 569.775911] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.775911] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] raise e [ 569.775911] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.775911] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] nwinfo = self.network_api.allocate_for_instance( [ 569.775911] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 569.775911] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] created_port_ids = self._update_ports_for_instance( [ 569.776497] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 569.776497] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] with excutils.save_and_reraise_exception(): [ 569.776497] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.776497] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] self.force_reraise() [ 569.776497] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.776497] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] raise self.value [ 569.776497] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 569.776497] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] updated_port = self._update_port( [ 569.776497] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.776497] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] _ensure_no_port_binding_failure(port) [ 569.776497] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.776497] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] raise exception.PortBindingFailed(port_id=port['id']) [ 569.776846] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] nova.exception.PortBindingFailed: Binding failed for port 3b92f861-eeb4-4364-9ab4-21f10dce1713, please check neutron logs for more information. [ 569.776846] env[61970]: ERROR nova.compute.manager [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] [ 569.776846] env[61970]: DEBUG nova.compute.utils [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Binding failed for port 3b92f861-eeb4-4364-9ab4-21f10dce1713, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 569.781586] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.513s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.781586] env[61970]: INFO nova.compute.claims [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 569.784262] env[61970]: DEBUG nova.compute.manager [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Build of instance bce7d001-2885-409c-b455-4c8a3ce8eb9e was re-scheduled: Binding failed for port 3b92f861-eeb4-4364-9ab4-21f10dce1713, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 569.784587] env[61970]: DEBUG nova.compute.manager [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 569.786035] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Acquiring lock "refresh_cache-bce7d001-2885-409c-b455-4c8a3ce8eb9e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.786035] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Acquired lock "refresh_cache-bce7d001-2885-409c-b455-4c8a3ce8eb9e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.786035] env[61970]: DEBUG nova.network.neutron [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 570.308894] env[61970]: DEBUG nova.network.neutron [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.360381] env[61970]: DEBUG nova.network.neutron [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.862189] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Releasing lock "refresh_cache-bce7d001-2885-409c-b455-4c8a3ce8eb9e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.863096] env[61970]: DEBUG nova.compute.manager [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 570.863096] env[61970]: DEBUG nova.compute.manager [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 570.863096] env[61970]: DEBUG nova.network.neutron [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 570.898276] env[61970]: DEBUG nova.network.neutron [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.158487] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb360fdf-8dfd-4245-b8fc-f4f8d5572811 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.170182] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d549ce1-6949-48c2-a688-54e4d44f9618 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.208933] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f008ae9c-4ed9-48fe-a484-9a0ba78812c1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.217008] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649db11f-df98-4311-ba0d-e8465177b9a8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.232473] env[61970]: DEBUG nova.compute.provider_tree [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.259502] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Acquiring lock "26f6843a-2548-47e0-9dc9-bc53eda32b60" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.259502] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Lock "26f6843a-2548-47e0-9dc9-bc53eda32b60" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.400323] env[61970]: DEBUG nova.network.neutron [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.735974] env[61970]: DEBUG nova.scheduler.client.report [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 571.905725] env[61970]: INFO nova.compute.manager [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: bce7d001-2885-409c-b455-4c8a3ce8eb9e] Took 1.04 seconds to deallocate network for instance. [ 572.243191] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.243697] env[61970]: DEBUG nova.compute.manager [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 572.248029] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.649s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.248115] env[61970]: INFO nova.compute.claims [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 572.754706] env[61970]: DEBUG nova.compute.utils [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 572.760437] env[61970]: DEBUG nova.compute.manager [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 572.760437] env[61970]: DEBUG nova.network.neutron [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 572.838782] env[61970]: DEBUG nova.policy [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7dc5afc9cf54d549141bd404ee6a88e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7113b6e5c93341dba9dd0b1ef3206aeb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 572.952894] env[61970]: INFO nova.scheduler.client.report [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Deleted allocations for instance bce7d001-2885-409c-b455-4c8a3ce8eb9e [ 573.261610] env[61970]: DEBUG nova.compute.manager [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 573.445063] env[61970]: DEBUG nova.network.neutron [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Successfully created port: 69991011-5146-47f0-a8a2-794791aa05ec {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 573.461343] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b1567fec-0f57-4c7e-9f14-f61ac8cf3569 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Lock "bce7d001-2885-409c-b455-4c8a3ce8eb9e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.671s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.610498] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad50c61-833c-4c2f-9f85-da99e8782ed9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.618818] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb4c7cca-93d1-4f86-af6b-3dbaecba760f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.651070] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b11a27c6-e479-4c80-9f17-1ffdff70eb3d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.658939] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a24bfe04-44fc-4e8d-8000-05800c8dd11f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.674944] env[61970]: DEBUG nova.compute.provider_tree [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 573.963969] env[61970]: DEBUG nova.compute.manager [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 574.177964] env[61970]: DEBUG nova.scheduler.client.report [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 574.279916] env[61970]: DEBUG nova.compute.manager [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 574.314389] env[61970]: DEBUG nova.virt.hardware [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 574.314713] env[61970]: DEBUG nova.virt.hardware [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 574.314889] env[61970]: DEBUG nova.virt.hardware [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 574.315083] env[61970]: DEBUG nova.virt.hardware [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 574.315226] env[61970]: DEBUG nova.virt.hardware [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 574.315367] env[61970]: DEBUG nova.virt.hardware [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 574.315571] env[61970]: DEBUG nova.virt.hardware [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 574.315967] env[61970]: DEBUG nova.virt.hardware [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 574.316177] env[61970]: DEBUG nova.virt.hardware [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 574.316344] env[61970]: DEBUG nova.virt.hardware [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 574.316509] env[61970]: DEBUG nova.virt.hardware [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 574.317435] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9230a70-5b8a-4575-99e1-042b2558431c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.326313] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c826796-24a6-4d70-8d8a-76c1d675e019 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.490245] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.688886] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.442s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.689661] env[61970]: DEBUG nova.compute.manager [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 574.694340] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.691s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.202398] env[61970]: DEBUG nova.compute.utils [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 575.208895] env[61970]: DEBUG nova.compute.manager [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 575.208895] env[61970]: DEBUG nova.network.neutron [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 575.289048] env[61970]: ERROR nova.compute.manager [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 69991011-5146-47f0-a8a2-794791aa05ec, please check neutron logs for more information. [ 575.289048] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 575.289048] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.289048] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 575.289048] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.289048] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 575.289048] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.289048] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 575.289048] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.289048] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 575.289048] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.289048] env[61970]: ERROR nova.compute.manager raise self.value [ 575.289048] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.289048] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 575.289048] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.289048] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 575.289524] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.289524] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 575.289524] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 69991011-5146-47f0-a8a2-794791aa05ec, please check neutron logs for more information. [ 575.289524] env[61970]: ERROR nova.compute.manager [ 575.289524] env[61970]: Traceback (most recent call last): [ 575.289524] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 575.289524] env[61970]: listener.cb(fileno) [ 575.289524] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.289524] env[61970]: result = function(*args, **kwargs) [ 575.289524] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 575.289524] env[61970]: return func(*args, **kwargs) [ 575.289524] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.289524] env[61970]: raise e [ 575.289524] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.289524] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 575.289524] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.289524] env[61970]: created_port_ids = self._update_ports_for_instance( [ 575.289524] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.289524] env[61970]: with excutils.save_and_reraise_exception(): [ 575.289524] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.289524] env[61970]: self.force_reraise() [ 575.289524] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.289524] env[61970]: raise self.value [ 575.289524] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.289524] env[61970]: updated_port = self._update_port( [ 575.289524] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.289524] env[61970]: _ensure_no_port_binding_failure(port) [ 575.289524] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.289524] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 575.290329] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 69991011-5146-47f0-a8a2-794791aa05ec, please check neutron logs for more information. [ 575.290329] env[61970]: Removing descriptor: 15 [ 575.290424] env[61970]: ERROR nova.compute.manager [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 69991011-5146-47f0-a8a2-794791aa05ec, please check neutron logs for more information. [ 575.290424] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Traceback (most recent call last): [ 575.290424] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 575.290424] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] yield resources [ 575.290424] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 575.290424] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] self.driver.spawn(context, instance, image_meta, [ 575.290424] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 575.290424] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.290424] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.290424] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] vm_ref = self.build_virtual_machine(instance, [ 575.290424] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.290793] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.290793] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.290793] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] for vif in network_info: [ 575.290793] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 575.290793] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] return self._sync_wrapper(fn, *args, **kwargs) [ 575.290793] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 575.290793] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] self.wait() [ 575.290793] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 575.290793] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] self[:] = self._gt.wait() [ 575.290793] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.290793] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] return self._exit_event.wait() [ 575.290793] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.290793] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] result = hub.switch() [ 575.291217] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.291217] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] return self.greenlet.switch() [ 575.291217] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.291217] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] result = function(*args, **kwargs) [ 575.291217] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 575.291217] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] return func(*args, **kwargs) [ 575.291217] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.291217] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] raise e [ 575.291217] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.291217] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] nwinfo = self.network_api.allocate_for_instance( [ 575.291217] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.291217] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] created_port_ids = self._update_ports_for_instance( [ 575.291217] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.291644] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] with excutils.save_and_reraise_exception(): [ 575.291644] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.291644] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] self.force_reraise() [ 575.291644] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.291644] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] raise self.value [ 575.291644] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.291644] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] updated_port = self._update_port( [ 575.291644] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.291644] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] _ensure_no_port_binding_failure(port) [ 575.291644] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.291644] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] raise exception.PortBindingFailed(port_id=port['id']) [ 575.291644] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] nova.exception.PortBindingFailed: Binding failed for port 69991011-5146-47f0-a8a2-794791aa05ec, please check neutron logs for more information. [ 575.291644] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] [ 575.292130] env[61970]: INFO nova.compute.manager [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Terminating instance [ 575.295263] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Acquiring lock "refresh_cache-39bc8558-ce7f-41e9-b913-9b628a220ca7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.295408] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Acquired lock "refresh_cache-39bc8558-ce7f-41e9-b913-9b628a220ca7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.295581] env[61970]: DEBUG nova.network.neutron [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 575.404338] env[61970]: DEBUG nova.policy [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac32ffb1d55848b69da08686932a8535', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9f43c08f5624dbb9d52bad90a8545aa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 575.578938] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee7954c2-2dd5-4a8a-b7d2-7bb72a863692 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.593097] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95196362-30b9-41db-9106-16a8e05dcf37 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.632900] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed3fbb4-b0fb-4fcf-99a5-ca4018910a89 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.644107] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-581b4c59-550a-4908-bf19-cc5a341a8d7e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.662415] env[61970]: DEBUG nova.compute.provider_tree [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 575.708777] env[61970]: DEBUG nova.compute.manager [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 575.714536] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "2933b91a-ab88-4f0f-bd63-3752cdc763de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.714753] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "2933b91a-ab88-4f0f-bd63-3752cdc763de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.832205] env[61970]: DEBUG nova.network.neutron [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.888227] env[61970]: DEBUG nova.compute.manager [req-443a6dd6-3737-4b74-866a-213b99f5ecd6 req-645e8eb6-432c-4a36-b0e0-21e7a5f42e3d service nova] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Received event network-changed-69991011-5146-47f0-a8a2-794791aa05ec {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 575.889858] env[61970]: DEBUG nova.compute.manager [req-443a6dd6-3737-4b74-866a-213b99f5ecd6 req-645e8eb6-432c-4a36-b0e0-21e7a5f42e3d service nova] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Refreshing instance network info cache due to event network-changed-69991011-5146-47f0-a8a2-794791aa05ec. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 575.889858] env[61970]: DEBUG oslo_concurrency.lockutils [req-443a6dd6-3737-4b74-866a-213b99f5ecd6 req-645e8eb6-432c-4a36-b0e0-21e7a5f42e3d service nova] Acquiring lock "refresh_cache-39bc8558-ce7f-41e9-b913-9b628a220ca7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.166429] env[61970]: DEBUG nova.scheduler.client.report [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 576.287876] env[61970]: DEBUG nova.network.neutron [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.673299] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.979s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.673846] env[61970]: ERROR nova.compute.manager [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port aae58430-6a65-4482-befa-f615cc7055d5, please check neutron logs for more information. [ 576.673846] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Traceback (most recent call last): [ 576.673846] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 576.673846] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] self.driver.spawn(context, instance, image_meta, [ 576.673846] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 576.673846] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 576.673846] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 576.673846] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] vm_ref = self.build_virtual_machine(instance, [ 576.673846] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 576.673846] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] vif_infos = vmwarevif.get_vif_info(self._session, [ 576.673846] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 576.674503] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] for vif in network_info: [ 576.674503] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 576.674503] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] return self._sync_wrapper(fn, *args, **kwargs) [ 576.674503] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 576.674503] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] self.wait() [ 576.674503] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 576.674503] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] self[:] = self._gt.wait() [ 576.674503] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 576.674503] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] return self._exit_event.wait() [ 576.674503] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 576.674503] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] result = hub.switch() [ 576.674503] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 576.674503] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] return self.greenlet.switch() [ 576.674910] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.674910] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] result = function(*args, **kwargs) [ 576.674910] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 576.674910] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] return func(*args, **kwargs) [ 576.674910] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 576.674910] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] raise e [ 576.674910] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 576.674910] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] nwinfo = self.network_api.allocate_for_instance( [ 576.674910] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 576.674910] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] created_port_ids = self._update_ports_for_instance( [ 576.674910] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 576.674910] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] with excutils.save_and_reraise_exception(): [ 576.674910] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.675279] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] self.force_reraise() [ 576.675279] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.675279] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] raise self.value [ 576.675279] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 576.675279] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] updated_port = self._update_port( [ 576.675279] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.675279] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] _ensure_no_port_binding_failure(port) [ 576.675279] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.675279] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] raise exception.PortBindingFailed(port_id=port['id']) [ 576.675279] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] nova.exception.PortBindingFailed: Binding failed for port aae58430-6a65-4482-befa-f615cc7055d5, please check neutron logs for more information. [ 576.675279] env[61970]: ERROR nova.compute.manager [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] [ 576.675616] env[61970]: DEBUG nova.compute.utils [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Binding failed for port aae58430-6a65-4482-befa-f615cc7055d5, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 576.678228] env[61970]: DEBUG nova.compute.manager [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Build of instance 36c69772-6e47-4c00-a582-b46d233e13a1 was re-scheduled: Binding failed for port aae58430-6a65-4482-befa-f615cc7055d5, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 576.678228] env[61970]: DEBUG nova.compute.manager [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 576.678228] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Acquiring lock "refresh_cache-36c69772-6e47-4c00-a582-b46d233e13a1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.678228] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Acquired lock "refresh_cache-36c69772-6e47-4c00-a582-b46d233e13a1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.678535] env[61970]: DEBUG nova.network.neutron [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 576.678535] env[61970]: DEBUG oslo_concurrency.lockutils [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.840s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.680549] env[61970]: INFO nova.compute.claims [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 576.722317] env[61970]: DEBUG nova.compute.manager [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 576.748827] env[61970]: DEBUG nova.virt.hardware [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 576.749082] env[61970]: DEBUG nova.virt.hardware [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 576.749242] env[61970]: DEBUG nova.virt.hardware [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 576.749424] env[61970]: DEBUG nova.virt.hardware [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 576.749700] env[61970]: DEBUG nova.virt.hardware [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 576.749700] env[61970]: DEBUG nova.virt.hardware [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 576.749903] env[61970]: DEBUG nova.virt.hardware [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 576.750107] env[61970]: DEBUG nova.virt.hardware [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 576.750286] env[61970]: DEBUG nova.virt.hardware [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 576.750468] env[61970]: DEBUG nova.virt.hardware [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 576.750641] env[61970]: DEBUG nova.virt.hardware [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 576.752086] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-901bbf9b-f37a-49e7-8f09-d00cba19e94f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.761989] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-959bd772-aef4-4ae0-b048-49506b4b7c76 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.794709] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Releasing lock "refresh_cache-39bc8558-ce7f-41e9-b913-9b628a220ca7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.794709] env[61970]: DEBUG nova.compute.manager [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 576.794709] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 576.794709] env[61970]: DEBUG oslo_concurrency.lockutils [req-443a6dd6-3737-4b74-866a-213b99f5ecd6 req-645e8eb6-432c-4a36-b0e0-21e7a5f42e3d service nova] Acquired lock "refresh_cache-39bc8558-ce7f-41e9-b913-9b628a220ca7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.794709] env[61970]: DEBUG nova.network.neutron [req-443a6dd6-3737-4b74-866a-213b99f5ecd6 req-645e8eb6-432c-4a36-b0e0-21e7a5f42e3d service nova] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Refreshing network info cache for port 69991011-5146-47f0-a8a2-794791aa05ec {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 576.794934] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-43a5fe30-ae02-4b40-9168-49bf84c42f17 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.807167] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75ae8abe-90f3-465c-8b15-c38eb611e506 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.832733] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 39bc8558-ce7f-41e9-b913-9b628a220ca7 could not be found. [ 576.832945] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 576.833133] env[61970]: INFO nova.compute.manager [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 576.833363] env[61970]: DEBUG oslo.service.loopingcall [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 576.833573] env[61970]: DEBUG nova.compute.manager [-] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 576.835031] env[61970]: DEBUG nova.network.neutron [-] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 576.843811] env[61970]: DEBUG nova.network.neutron [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Successfully created port: 09574426-d427-42af-9a8a-a9d325d1154d {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 576.889324] env[61970]: DEBUG nova.network.neutron [-] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.109010] env[61970]: DEBUG nova.compute.manager [None req-8d8f71b5-94ec-4ff0-8313-03e7b37e1778 tempest-ServerDiagnosticsV248Test-831209024 tempest-ServerDiagnosticsV248Test-831209024-project-admin] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 577.110501] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1abbed9-f2e1-4cd5-97df-3ce159778344 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.120862] env[61970]: INFO nova.compute.manager [None req-8d8f71b5-94ec-4ff0-8313-03e7b37e1778 tempest-ServerDiagnosticsV248Test-831209024 tempest-ServerDiagnosticsV248Test-831209024-project-admin] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Retrieving diagnostics [ 577.122012] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e066b9a-3854-4111-ad37-239f12e48eb2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.246155] env[61970]: DEBUG nova.network.neutron [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.381107] env[61970]: DEBUG nova.network.neutron [req-443a6dd6-3737-4b74-866a-213b99f5ecd6 req-645e8eb6-432c-4a36-b0e0-21e7a5f42e3d service nova] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.392494] env[61970]: DEBUG nova.network.neutron [-] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.527619] env[61970]: DEBUG nova.network.neutron [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.855548] env[61970]: DEBUG nova.network.neutron [req-443a6dd6-3737-4b74-866a-213b99f5ecd6 req-645e8eb6-432c-4a36-b0e0-21e7a5f42e3d service nova] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.896771] env[61970]: INFO nova.compute.manager [-] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Took 1.06 seconds to deallocate network for instance. [ 577.901791] env[61970]: DEBUG nova.compute.claims [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 577.901950] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.031993] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Releasing lock "refresh_cache-36c69772-6e47-4c00-a582-b46d233e13a1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.034908] env[61970]: DEBUG nova.compute.manager [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 578.034908] env[61970]: DEBUG nova.compute.manager [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 578.034908] env[61970]: DEBUG nova.network.neutron [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 578.096968] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b066e3d-2166-43fe-9cda-5999f833d0e8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.102578] env[61970]: DEBUG nova.network.neutron [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.111167] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa5ccaf0-ede1-4598-b420-74a915beda8d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.152686] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f527e6-059f-41d5-b5a5-fbbcded27327 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.161195] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8df064f9-0915-448f-825a-3cfe7b6b712b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.184483] env[61970]: DEBUG nova.compute.provider_tree [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 578.358821] env[61970]: DEBUG oslo_concurrency.lockutils [req-443a6dd6-3737-4b74-866a-213b99f5ecd6 req-645e8eb6-432c-4a36-b0e0-21e7a5f42e3d service nova] Releasing lock "refresh_cache-39bc8558-ce7f-41e9-b913-9b628a220ca7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.414610] env[61970]: DEBUG nova.compute.manager [req-8b3f8a4b-301b-4f54-94b4-67e4592fb0a7 req-573dc35d-b05a-480b-9faf-16c9847149a2 service nova] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Received event network-vif-deleted-69991011-5146-47f0-a8a2-794791aa05ec {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 578.606177] env[61970]: DEBUG nova.network.neutron [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.688503] env[61970]: DEBUG nova.scheduler.client.report [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 578.747360] env[61970]: DEBUG oslo_concurrency.lockutils [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Acquiring lock "8881014c-8505-4e75-8c31-e25b6a8d2cd4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.749235] env[61970]: DEBUG oslo_concurrency.lockutils [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Lock "8881014c-8505-4e75-8c31-e25b6a8d2cd4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.749442] env[61970]: DEBUG oslo_concurrency.lockutils [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Acquiring lock "8881014c-8505-4e75-8c31-e25b6a8d2cd4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.749650] env[61970]: DEBUG oslo_concurrency.lockutils [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Lock "8881014c-8505-4e75-8c31-e25b6a8d2cd4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.749888] env[61970]: DEBUG oslo_concurrency.lockutils [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Lock "8881014c-8505-4e75-8c31-e25b6a8d2cd4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.752047] env[61970]: INFO nova.compute.manager [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Terminating instance [ 578.753934] env[61970]: DEBUG oslo_concurrency.lockutils [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Acquiring lock "refresh_cache-8881014c-8505-4e75-8c31-e25b6a8d2cd4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.754049] env[61970]: DEBUG oslo_concurrency.lockutils [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Acquired lock "refresh_cache-8881014c-8505-4e75-8c31-e25b6a8d2cd4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.754398] env[61970]: DEBUG nova.network.neutron [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 579.110341] env[61970]: INFO nova.compute.manager [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] [instance: 36c69772-6e47-4c00-a582-b46d233e13a1] Took 1.08 seconds to deallocate network for instance. [ 579.200581] env[61970]: DEBUG oslo_concurrency.lockutils [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.520s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.201124] env[61970]: DEBUG nova.compute.manager [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 579.204603] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.628s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.206044] env[61970]: INFO nova.compute.claims [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 579.311142] env[61970]: DEBUG nova.network.neutron [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.557581] env[61970]: DEBUG nova.network.neutron [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.708786] env[61970]: DEBUG nova.compute.utils [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 579.709654] env[61970]: DEBUG nova.compute.manager [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 579.710041] env[61970]: DEBUG nova.network.neutron [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 580.000121] env[61970]: DEBUG nova.policy [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac32ffb1d55848b69da08686932a8535', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9f43c08f5624dbb9d52bad90a8545aa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 580.063264] env[61970]: DEBUG oslo_concurrency.lockutils [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Releasing lock "refresh_cache-8881014c-8505-4e75-8c31-e25b6a8d2cd4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.063264] env[61970]: DEBUG nova.compute.manager [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 580.063264] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 580.063264] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85377ba-344d-4184-9c5d-7d727abfb8b9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.074122] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 580.074122] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08c8f6bc-a944-4482-87af-6f98284b0bba {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.079866] env[61970]: DEBUG oslo_vmware.api [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Waiting for the task: (returnval){ [ 580.079866] env[61970]: value = "task-1355495" [ 580.079866] env[61970]: _type = "Task" [ 580.079866] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 580.090624] env[61970]: DEBUG oslo_vmware.api [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': task-1355495, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.155166] env[61970]: INFO nova.scheduler.client.report [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Deleted allocations for instance 36c69772-6e47-4c00-a582-b46d233e13a1 [ 580.214343] env[61970]: DEBUG nova.compute.manager [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 580.391661] env[61970]: DEBUG oslo_concurrency.lockutils [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Acquiring lock "de9fde95-3ad2-414a-b561-fe4ccc1dc9ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.391661] env[61970]: DEBUG oslo_concurrency.lockutils [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Lock "de9fde95-3ad2-414a-b561-fe4ccc1dc9ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.569826] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbba5281-9d02-42e9-a207-f4ef5af38e56 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.578811] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb1ea5a-dc53-4ea4-9cf7-1270b1043d8d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.617747] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d70966f-5ff1-4858-a94d-9cac12b43dae {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.620443] env[61970]: DEBUG oslo_vmware.api [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': task-1355495, 'name': PowerOffVM_Task, 'duration_secs': 0.153465} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 580.620875] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 580.620875] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 580.622131] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-57399765-38bb-44ea-aa2e-98748b2126a6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.626475] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e709c31-fae0-4c1a-87cb-f8d882055d57 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.640248] env[61970]: DEBUG nova.compute.provider_tree [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 580.645157] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 580.645737] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 580.645737] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Deleting the datastore file [datastore2] 8881014c-8505-4e75-8c31-e25b6a8d2cd4 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 580.645737] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a087400e-18b8-4500-b7d2-2bd8ff92dfb6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.652811] env[61970]: DEBUG oslo_vmware.api [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Waiting for the task: (returnval){ [ 580.652811] env[61970]: value = "task-1355497" [ 580.652811] env[61970]: _type = "Task" [ 580.652811] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 580.663983] env[61970]: DEBUG oslo_vmware.api [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': task-1355497, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.667248] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a62faba2-c64c-4c39-a2ac-533b5a806baf tempest-ServerAddressesTestJSON-366551292 tempest-ServerAddressesTestJSON-366551292-project-member] Lock "36c69772-6e47-4c00-a582-b46d233e13a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.983s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.143945] env[61970]: DEBUG nova.scheduler.client.report [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 581.166605] env[61970]: DEBUG oslo_vmware.api [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Task: {'id': task-1355497, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.111018} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.166605] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 581.166605] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 581.168962] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 581.168962] env[61970]: INFO nova.compute.manager [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Took 1.11 seconds to destroy the instance on the hypervisor. [ 581.168962] env[61970]: DEBUG oslo.service.loopingcall [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 581.168962] env[61970]: DEBUG nova.compute.manager [-] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 581.168962] env[61970]: DEBUG nova.network.neutron [-] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 581.172690] env[61970]: ERROR nova.compute.manager [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 09574426-d427-42af-9a8a-a9d325d1154d, please check neutron logs for more information. [ 581.172690] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 581.172690] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.172690] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 581.172690] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.172690] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 581.172690] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.172690] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 581.172690] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.172690] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 581.172690] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.172690] env[61970]: ERROR nova.compute.manager raise self.value [ 581.172690] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.172690] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 581.172690] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.172690] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 581.173193] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.173193] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 581.173193] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 09574426-d427-42af-9a8a-a9d325d1154d, please check neutron logs for more information. [ 581.173193] env[61970]: ERROR nova.compute.manager [ 581.173193] env[61970]: Traceback (most recent call last): [ 581.173193] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 581.173193] env[61970]: listener.cb(fileno) [ 581.173193] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.173193] env[61970]: result = function(*args, **kwargs) [ 581.173193] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 581.173193] env[61970]: return func(*args, **kwargs) [ 581.173193] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 581.173193] env[61970]: raise e [ 581.173193] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.173193] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 581.173193] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.173193] env[61970]: created_port_ids = self._update_ports_for_instance( [ 581.173193] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.173193] env[61970]: with excutils.save_and_reraise_exception(): [ 581.173193] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.173193] env[61970]: self.force_reraise() [ 581.173193] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.173193] env[61970]: raise self.value [ 581.173193] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.173193] env[61970]: updated_port = self._update_port( [ 581.173193] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.173193] env[61970]: _ensure_no_port_binding_failure(port) [ 581.173193] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.173193] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 581.174089] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 09574426-d427-42af-9a8a-a9d325d1154d, please check neutron logs for more information. [ 581.174089] env[61970]: Removing descriptor: 17 [ 581.174089] env[61970]: ERROR nova.compute.manager [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 09574426-d427-42af-9a8a-a9d325d1154d, please check neutron logs for more information. [ 581.174089] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Traceback (most recent call last): [ 581.174089] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 581.174089] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] yield resources [ 581.174089] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 581.174089] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] self.driver.spawn(context, instance, image_meta, [ 581.174089] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 581.174089] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 581.174089] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 581.174089] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] vm_ref = self.build_virtual_machine(instance, [ 581.174453] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 581.174453] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] vif_infos = vmwarevif.get_vif_info(self._session, [ 581.174453] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 581.174453] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] for vif in network_info: [ 581.174453] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 581.174453] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] return self._sync_wrapper(fn, *args, **kwargs) [ 581.174453] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 581.174453] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] self.wait() [ 581.174453] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 581.174453] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] self[:] = self._gt.wait() [ 581.174453] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 581.174453] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] return self._exit_event.wait() [ 581.174453] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 581.175291] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] result = hub.switch() [ 581.175291] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 581.175291] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] return self.greenlet.switch() [ 581.175291] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.175291] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] result = function(*args, **kwargs) [ 581.175291] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 581.175291] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] return func(*args, **kwargs) [ 581.175291] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 581.175291] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] raise e [ 581.175291] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.175291] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] nwinfo = self.network_api.allocate_for_instance( [ 581.175291] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.175291] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] created_port_ids = self._update_ports_for_instance( [ 581.175691] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.175691] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] with excutils.save_and_reraise_exception(): [ 581.175691] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.175691] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] self.force_reraise() [ 581.175691] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.175691] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] raise self.value [ 581.175691] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.175691] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] updated_port = self._update_port( [ 581.175691] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.175691] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] _ensure_no_port_binding_failure(port) [ 581.175691] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.175691] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] raise exception.PortBindingFailed(port_id=port['id']) [ 581.176131] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] nova.exception.PortBindingFailed: Binding failed for port 09574426-d427-42af-9a8a-a9d325d1154d, please check neutron logs for more information. [ 581.176131] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] [ 581.176131] env[61970]: INFO nova.compute.manager [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Terminating instance [ 581.179290] env[61970]: DEBUG nova.compute.manager [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 581.180346] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Acquiring lock "refresh_cache-17f41a87-4fd0-4866-905a-2aa3d2e9691b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.180650] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Acquired lock "refresh_cache-17f41a87-4fd0-4866-905a-2aa3d2e9691b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.180925] env[61970]: DEBUG nova.network.neutron [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 581.209749] env[61970]: DEBUG nova.network.neutron [-] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 581.225381] env[61970]: DEBUG nova.compute.manager [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 581.269241] env[61970]: DEBUG nova.virt.hardware [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 581.269865] env[61970]: DEBUG nova.virt.hardware [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 581.269865] env[61970]: DEBUG nova.virt.hardware [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 581.269865] env[61970]: DEBUG nova.virt.hardware [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 581.270180] env[61970]: DEBUG nova.virt.hardware [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 581.270407] env[61970]: DEBUG nova.virt.hardware [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 581.270823] env[61970]: DEBUG nova.virt.hardware [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 581.271187] env[61970]: DEBUG nova.virt.hardware [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 581.271379] env[61970]: DEBUG nova.virt.hardware [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 581.271546] env[61970]: DEBUG nova.virt.hardware [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 581.271716] env[61970]: DEBUG nova.virt.hardware [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 581.272966] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b0b251-960d-4f44-a097-eafa014597ec {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.282892] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ca4e0a-f750-4e42-963b-b8a80f28a117 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.476632] env[61970]: DEBUG nova.network.neutron [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Successfully created port: abc054de-ef96-40b1-8b51-30bf5252429d {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 581.649700] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.650555] env[61970]: DEBUG nova.compute.manager [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 581.656165] env[61970]: DEBUG oslo_concurrency.lockutils [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.601s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.706332] env[61970]: DEBUG oslo_concurrency.lockutils [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.714366] env[61970]: DEBUG nova.network.neutron [-] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.756184] env[61970]: DEBUG nova.network.neutron [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 581.760727] env[61970]: DEBUG nova.compute.manager [req-4b4704eb-81b0-4be9-8344-c1ed2a937134 req-362328f0-35a2-44ae-9e22-6b94979e91fb service nova] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Received event network-changed-09574426-d427-42af-9a8a-a9d325d1154d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 581.760727] env[61970]: DEBUG nova.compute.manager [req-4b4704eb-81b0-4be9-8344-c1ed2a937134 req-362328f0-35a2-44ae-9e22-6b94979e91fb service nova] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Refreshing instance network info cache due to event network-changed-09574426-d427-42af-9a8a-a9d325d1154d. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 581.760727] env[61970]: DEBUG oslo_concurrency.lockutils [req-4b4704eb-81b0-4be9-8344-c1ed2a937134 req-362328f0-35a2-44ae-9e22-6b94979e91fb service nova] Acquiring lock "refresh_cache-17f41a87-4fd0-4866-905a-2aa3d2e9691b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.015076] env[61970]: DEBUG nova.network.neutron [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.067778] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Acquiring lock "293b6863-510c-493e-8c9a-af53ad80fc89" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.067934] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Lock "293b6863-510c-493e-8c9a-af53ad80fc89" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.164939] env[61970]: DEBUG nova.compute.utils [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 582.170702] env[61970]: DEBUG nova.compute.manager [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 582.172105] env[61970]: DEBUG nova.network.neutron [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 582.218982] env[61970]: INFO nova.compute.manager [-] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Took 1.05 seconds to deallocate network for instance. [ 582.523091] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Releasing lock "refresh_cache-17f41a87-4fd0-4866-905a-2aa3d2e9691b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.523091] env[61970]: DEBUG nova.compute.manager [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 582.523091] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 582.527489] env[61970]: DEBUG oslo_concurrency.lockutils [req-4b4704eb-81b0-4be9-8344-c1ed2a937134 req-362328f0-35a2-44ae-9e22-6b94979e91fb service nova] Acquired lock "refresh_cache-17f41a87-4fd0-4866-905a-2aa3d2e9691b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.527663] env[61970]: DEBUG nova.network.neutron [req-4b4704eb-81b0-4be9-8344-c1ed2a937134 req-362328f0-35a2-44ae-9e22-6b94979e91fb service nova] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Refreshing network info cache for port 09574426-d427-42af-9a8a-a9d325d1154d {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 582.528883] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2677337c-e68b-4661-80ae-c871302c88a0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.540160] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec3c0dc-b3b4-4972-9be7-653da77d68b5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.566889] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 17f41a87-4fd0-4866-905a-2aa3d2e9691b could not be found. [ 582.566889] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 582.567186] env[61970]: INFO nova.compute.manager [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 582.567336] env[61970]: DEBUG oslo.service.loopingcall [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 582.570534] env[61970]: DEBUG nova.compute.manager [-] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 582.570627] env[61970]: DEBUG nova.network.neutron [-] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 582.573577] env[61970]: DEBUG nova.policy [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6098a286ae994a5287aeb40039623cf8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '578451cfc36f453997d89ab8a507132a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 582.599606] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc67d991-9681-404b-a7c9-f39af0a684c1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.607937] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dda5400-3f80-43e2-bc18-190d14db224a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.638994] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429db2ca-9fbc-4305-b529-58d93ec92ae8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.642812] env[61970]: DEBUG nova.network.neutron [-] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.648205] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624b689d-3063-4dbf-baef-e21726399ede {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.662254] env[61970]: DEBUG nova.compute.provider_tree [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 582.671687] env[61970]: DEBUG nova.compute.manager [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 582.729430] env[61970]: DEBUG oslo_concurrency.lockutils [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.083347] env[61970]: DEBUG nova.network.neutron [req-4b4704eb-81b0-4be9-8344-c1ed2a937134 req-362328f0-35a2-44ae-9e22-6b94979e91fb service nova] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.144766] env[61970]: DEBUG nova.network.neutron [-] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.165331] env[61970]: DEBUG nova.scheduler.client.report [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 583.394033] env[61970]: DEBUG nova.network.neutron [req-4b4704eb-81b0-4be9-8344-c1ed2a937134 req-362328f0-35a2-44ae-9e22-6b94979e91fb service nova] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.652941] env[61970]: INFO nova.compute.manager [-] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Took 1.08 seconds to deallocate network for instance. [ 583.658323] env[61970]: DEBUG nova.compute.claims [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 583.658516] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.674666] env[61970]: DEBUG oslo_concurrency.lockutils [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.020s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.674666] env[61970]: ERROR nova.compute.manager [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c09b9037-d99d-4c4e-950b-27276f22b702, please check neutron logs for more information. [ 583.674666] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Traceback (most recent call last): [ 583.674666] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 583.674666] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] self.driver.spawn(context, instance, image_meta, [ 583.674666] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 583.674666] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] self._vmops.spawn(context, instance, image_meta, injected_files, [ 583.674666] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 583.674666] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] vm_ref = self.build_virtual_machine(instance, [ 583.675210] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 583.675210] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] vif_infos = vmwarevif.get_vif_info(self._session, [ 583.675210] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 583.675210] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] for vif in network_info: [ 583.675210] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 583.675210] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] return self._sync_wrapper(fn, *args, **kwargs) [ 583.675210] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 583.675210] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] self.wait() [ 583.675210] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 583.675210] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] self[:] = self._gt.wait() [ 583.675210] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 583.675210] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] return self._exit_event.wait() [ 583.675210] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 583.675616] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] result = hub.switch() [ 583.675616] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 583.675616] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] return self.greenlet.switch() [ 583.675616] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.675616] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] result = function(*args, **kwargs) [ 583.675616] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 583.675616] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] return func(*args, **kwargs) [ 583.675616] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 583.675616] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] raise e [ 583.675616] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.675616] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] nwinfo = self.network_api.allocate_for_instance( [ 583.675616] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.675616] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] created_port_ids = self._update_ports_for_instance( [ 583.676071] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.676071] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] with excutils.save_and_reraise_exception(): [ 583.676071] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.676071] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] self.force_reraise() [ 583.676071] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.676071] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] raise self.value [ 583.676071] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.676071] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] updated_port = self._update_port( [ 583.676071] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.676071] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] _ensure_no_port_binding_failure(port) [ 583.676071] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.676071] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] raise exception.PortBindingFailed(port_id=port['id']) [ 583.676466] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] nova.exception.PortBindingFailed: Binding failed for port c09b9037-d99d-4c4e-950b-27276f22b702, please check neutron logs for more information. [ 583.676466] env[61970]: ERROR nova.compute.manager [instance: e3744562-c9ab-4902-880c-6f0153b8a183] [ 583.676466] env[61970]: DEBUG nova.compute.utils [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Binding failed for port c09b9037-d99d-4c4e-950b-27276f22b702, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 583.678176] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.021s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.678176] env[61970]: DEBUG nova.objects.instance [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Lazy-loading 'resources' on Instance uuid c4d3b145-84be-4d19-afc7-4c29fa193760 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 583.678473] env[61970]: DEBUG nova.compute.manager [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Build of instance e3744562-c9ab-4902-880c-6f0153b8a183 was re-scheduled: Binding failed for port c09b9037-d99d-4c4e-950b-27276f22b702, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 583.681017] env[61970]: DEBUG nova.compute.manager [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 583.681017] env[61970]: DEBUG oslo_concurrency.lockutils [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Acquiring lock "refresh_cache-e3744562-c9ab-4902-880c-6f0153b8a183" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.681017] env[61970]: DEBUG oslo_concurrency.lockutils [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Acquired lock "refresh_cache-e3744562-c9ab-4902-880c-6f0153b8a183" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.681017] env[61970]: DEBUG nova.network.neutron [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 583.685074] env[61970]: DEBUG nova.compute.manager [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 583.726223] env[61970]: DEBUG nova.virt.hardware [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 583.726461] env[61970]: DEBUG nova.virt.hardware [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 583.726620] env[61970]: DEBUG nova.virt.hardware [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 583.726798] env[61970]: DEBUG nova.virt.hardware [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 583.726928] env[61970]: DEBUG nova.virt.hardware [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 583.727901] env[61970]: DEBUG nova.virt.hardware [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 583.728226] env[61970]: DEBUG nova.virt.hardware [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 583.728415] env[61970]: DEBUG nova.virt.hardware [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 583.729581] env[61970]: DEBUG nova.virt.hardware [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 583.729581] env[61970]: DEBUG nova.virt.hardware [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 583.729581] env[61970]: DEBUG nova.virt.hardware [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 583.733570] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fdb12a3-d6c9-4240-93f9-d018da3d9f79 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.742020] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3752d4ca-2654-4341-8f6a-16c1bbafe2c5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.751209] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Acquiring lock "77460934-215e-4294-8eb3-a380dfbacddd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.751442] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Lock "77460934-215e-4294-8eb3-a380dfbacddd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.900019] env[61970]: DEBUG oslo_concurrency.lockutils [req-4b4704eb-81b0-4be9-8344-c1ed2a937134 req-362328f0-35a2-44ae-9e22-6b94979e91fb service nova] Releasing lock "refresh_cache-17f41a87-4fd0-4866-905a-2aa3d2e9691b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.974290] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "90cb79fb-aeb4-4c13-ab80-2787f67ff240" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.974662] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "90cb79fb-aeb4-4c13-ab80-2787f67ff240" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.256602] env[61970]: ERROR nova.compute.manager [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port abc054de-ef96-40b1-8b51-30bf5252429d, please check neutron logs for more information. [ 584.256602] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 584.256602] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.256602] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 584.256602] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 584.256602] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 584.256602] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 584.256602] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 584.256602] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.256602] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 584.256602] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.256602] env[61970]: ERROR nova.compute.manager raise self.value [ 584.256602] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 584.256602] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 584.256602] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.256602] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 584.257405] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.257405] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 584.257405] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port abc054de-ef96-40b1-8b51-30bf5252429d, please check neutron logs for more information. [ 584.257405] env[61970]: ERROR nova.compute.manager [ 584.257405] env[61970]: Traceback (most recent call last): [ 584.257405] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 584.257405] env[61970]: listener.cb(fileno) [ 584.257405] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.257405] env[61970]: result = function(*args, **kwargs) [ 584.257405] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 584.257405] env[61970]: return func(*args, **kwargs) [ 584.257405] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 584.257405] env[61970]: raise e [ 584.257405] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.257405] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 584.257405] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 584.257405] env[61970]: created_port_ids = self._update_ports_for_instance( [ 584.257405] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 584.257405] env[61970]: with excutils.save_and_reraise_exception(): [ 584.257405] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.257405] env[61970]: self.force_reraise() [ 584.257405] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.257405] env[61970]: raise self.value [ 584.257405] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 584.257405] env[61970]: updated_port = self._update_port( [ 584.257405] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.257405] env[61970]: _ensure_no_port_binding_failure(port) [ 584.257405] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.257405] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 584.258294] env[61970]: nova.exception.PortBindingFailed: Binding failed for port abc054de-ef96-40b1-8b51-30bf5252429d, please check neutron logs for more information. [ 584.258294] env[61970]: Removing descriptor: 20 [ 584.258294] env[61970]: ERROR nova.compute.manager [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port abc054de-ef96-40b1-8b51-30bf5252429d, please check neutron logs for more information. [ 584.258294] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Traceback (most recent call last): [ 584.258294] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 584.258294] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] yield resources [ 584.258294] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 584.258294] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] self.driver.spawn(context, instance, image_meta, [ 584.258294] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 584.258294] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 584.258294] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 584.258294] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] vm_ref = self.build_virtual_machine(instance, [ 584.258630] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 584.258630] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] vif_infos = vmwarevif.get_vif_info(self._session, [ 584.258630] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 584.258630] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] for vif in network_info: [ 584.258630] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 584.258630] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] return self._sync_wrapper(fn, *args, **kwargs) [ 584.258630] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 584.258630] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] self.wait() [ 584.258630] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 584.258630] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] self[:] = self._gt.wait() [ 584.258630] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 584.258630] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] return self._exit_event.wait() [ 584.258630] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 584.259018] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] result = hub.switch() [ 584.259018] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 584.259018] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] return self.greenlet.switch() [ 584.259018] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.259018] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] result = function(*args, **kwargs) [ 584.259018] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 584.259018] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] return func(*args, **kwargs) [ 584.259018] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 584.259018] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] raise e [ 584.259018] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.259018] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] nwinfo = self.network_api.allocate_for_instance( [ 584.259018] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 584.259018] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] created_port_ids = self._update_ports_for_instance( [ 584.259380] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 584.259380] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] with excutils.save_and_reraise_exception(): [ 584.259380] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.259380] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] self.force_reraise() [ 584.259380] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.259380] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] raise self.value [ 584.259380] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 584.259380] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] updated_port = self._update_port( [ 584.259380] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.259380] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] _ensure_no_port_binding_failure(port) [ 584.259380] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.259380] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] raise exception.PortBindingFailed(port_id=port['id']) [ 584.259726] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] nova.exception.PortBindingFailed: Binding failed for port abc054de-ef96-40b1-8b51-30bf5252429d, please check neutron logs for more information. [ 584.259726] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] [ 584.259726] env[61970]: INFO nova.compute.manager [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Terminating instance [ 584.261603] env[61970]: DEBUG nova.network.neutron [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.265009] env[61970]: DEBUG oslo_concurrency.lockutils [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Acquiring lock "refresh_cache-8bacc19d-2626-406a-a3fb-f1069674963c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.265572] env[61970]: DEBUG oslo_concurrency.lockutils [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Acquired lock "refresh_cache-8bacc19d-2626-406a-a3fb-f1069674963c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.265572] env[61970]: DEBUG nova.network.neutron [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 584.366331] env[61970]: DEBUG nova.network.neutron [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Successfully created port: 9ca8ea05-ac30-4e52-8de3-5ec6b39bdc8f {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 584.658750] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7040bfe3-3170-449e-b608-ff8a050aedb0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.667183] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51cab726-a395-46ee-8d76-a5e739e6442e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.705040] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a73488d0-9935-4533-8501-2652a75b8a69 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.714638] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ddc4bd-199b-414f-a915-478be886d089 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.729395] env[61970]: DEBUG nova.compute.provider_tree [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.744681] env[61970]: DEBUG nova.network.neutron [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.851441] env[61970]: DEBUG nova.network.neutron [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.858471] env[61970]: DEBUG nova.compute.manager [req-c9b39a11-0912-43b4-853e-897e199ffc3d req-b89185d0-c68d-4c22-a49c-172293661545 service nova] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Received event network-vif-deleted-09574426-d427-42af-9a8a-a9d325d1154d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 584.872802] env[61970]: DEBUG nova.compute.manager [req-a8cd4234-a3e0-4110-8035-bc605061a5e1 req-3657eca9-c5e1-4755-b0ba-b6a6c8aefca3 service nova] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Received event network-changed-abc054de-ef96-40b1-8b51-30bf5252429d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 584.872987] env[61970]: DEBUG nova.compute.manager [req-a8cd4234-a3e0-4110-8035-bc605061a5e1 req-3657eca9-c5e1-4755-b0ba-b6a6c8aefca3 service nova] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Refreshing instance network info cache due to event network-changed-abc054de-ef96-40b1-8b51-30bf5252429d. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 584.873183] env[61970]: DEBUG oslo_concurrency.lockutils [req-a8cd4234-a3e0-4110-8035-bc605061a5e1 req-3657eca9-c5e1-4755-b0ba-b6a6c8aefca3 service nova] Acquiring lock "refresh_cache-8bacc19d-2626-406a-a3fb-f1069674963c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.179951] env[61970]: DEBUG nova.network.neutron [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.236111] env[61970]: DEBUG nova.scheduler.client.report [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 585.247776] env[61970]: DEBUG oslo_concurrency.lockutils [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Releasing lock "refresh_cache-e3744562-c9ab-4902-880c-6f0153b8a183" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.248034] env[61970]: DEBUG nova.compute.manager [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 585.250018] env[61970]: DEBUG nova.compute.manager [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 585.250018] env[61970]: DEBUG nova.network.neutron [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 585.321065] env[61970]: DEBUG nova.network.neutron [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.686808] env[61970]: DEBUG oslo_concurrency.lockutils [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Releasing lock "refresh_cache-8bacc19d-2626-406a-a3fb-f1069674963c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.687472] env[61970]: DEBUG nova.compute.manager [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 585.687472] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 585.687811] env[61970]: DEBUG oslo_concurrency.lockutils [req-a8cd4234-a3e0-4110-8035-bc605061a5e1 req-3657eca9-c5e1-4755-b0ba-b6a6c8aefca3 service nova] Acquired lock "refresh_cache-8bacc19d-2626-406a-a3fb-f1069674963c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.687981] env[61970]: DEBUG nova.network.neutron [req-a8cd4234-a3e0-4110-8035-bc605061a5e1 req-3657eca9-c5e1-4755-b0ba-b6a6c8aefca3 service nova] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Refreshing network info cache for port abc054de-ef96-40b1-8b51-30bf5252429d {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 585.689094] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-78a5e6f5-3773-4d33-b508-2f2648f1523a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.699641] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc2f3a9-33cc-4dc6-bc3c-a6e144c32f18 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.727571] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8bacc19d-2626-406a-a3fb-f1069674963c could not be found. [ 585.727874] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 585.727997] env[61970]: INFO nova.compute.manager [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 585.730571] env[61970]: DEBUG oslo.service.loopingcall [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 585.730571] env[61970]: DEBUG nova.compute.manager [-] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 585.730571] env[61970]: DEBUG nova.network.neutron [-] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 585.747911] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.068s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.747911] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.403s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.755219] env[61970]: INFO nova.compute.claims [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 585.775802] env[61970]: INFO nova.scheduler.client.report [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Deleted allocations for instance c4d3b145-84be-4d19-afc7-4c29fa193760 [ 585.814654] env[61970]: DEBUG nova.network.neutron [-] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.826153] env[61970]: DEBUG nova.network.neutron [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.291471] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6fbe963-e2ac-435c-802b-50950cf1211f tempest-ServersAaction247Test-1596232376 tempest-ServersAaction247Test-1596232376-project-member] Lock "c4d3b145-84be-4d19-afc7-4c29fa193760" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.575s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.317409] env[61970]: DEBUG nova.network.neutron [-] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.330966] env[61970]: INFO nova.compute.manager [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] [instance: e3744562-c9ab-4902-880c-6f0153b8a183] Took 1.08 seconds to deallocate network for instance. [ 586.358029] env[61970]: DEBUG nova.network.neutron [req-a8cd4234-a3e0-4110-8035-bc605061a5e1 req-3657eca9-c5e1-4755-b0ba-b6a6c8aefca3 service nova] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.570583] env[61970]: DEBUG nova.network.neutron [req-a8cd4234-a3e0-4110-8035-bc605061a5e1 req-3657eca9-c5e1-4755-b0ba-b6a6c8aefca3 service nova] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.822639] env[61970]: INFO nova.compute.manager [-] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Took 1.09 seconds to deallocate network for instance. [ 586.827909] env[61970]: DEBUG nova.compute.claims [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 586.832155] env[61970]: DEBUG oslo_concurrency.lockutils [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.073961] env[61970]: DEBUG oslo_concurrency.lockutils [req-a8cd4234-a3e0-4110-8035-bc605061a5e1 req-3657eca9-c5e1-4755-b0ba-b6a6c8aefca3 service nova] Releasing lock "refresh_cache-8bacc19d-2626-406a-a3fb-f1069674963c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.227742] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f48d6b91-1373-411a-ad65-b955cd1fd24f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.234769] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a50cdc87-8d2c-4c46-a147-bd49cc53a285 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.268134] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249cc384-bfc4-41ba-aa6b-c517bace58f5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.279430] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a321bb-0b67-447d-886a-df7ee830495f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.298308] env[61970]: DEBUG nova.compute.provider_tree [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.388022] env[61970]: INFO nova.scheduler.client.report [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Deleted allocations for instance e3744562-c9ab-4902-880c-6f0153b8a183 [ 587.422418] env[61970]: DEBUG nova.compute.manager [req-2b85ff41-dbfb-47d1-826b-03e0b9cddf98 req-6e6596cb-dfba-4ad1-a23e-4d98e28d01a2 service nova] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Received event network-vif-deleted-abc054de-ef96-40b1-8b51-30bf5252429d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 587.805073] env[61970]: DEBUG nova.scheduler.client.report [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 587.902621] env[61970]: DEBUG oslo_concurrency.lockutils [None req-363635b5-60cd-46d5-a7f8-53181b6ff2f5 tempest-ServerActionsTestJSON-1768007552 tempest-ServerActionsTestJSON-1768007552-project-member] Lock "e3744562-c9ab-4902-880c-6f0153b8a183" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.015s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.320128] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.570s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.320128] env[61970]: DEBUG nova.compute.manager [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 588.323727] env[61970]: DEBUG oslo_concurrency.lockutils [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.711s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.323727] env[61970]: INFO nova.compute.claims [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 588.409503] env[61970]: DEBUG nova.compute.manager [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 588.833484] env[61970]: DEBUG nova.compute.utils [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 588.843148] env[61970]: DEBUG nova.compute.manager [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 588.843148] env[61970]: DEBUG nova.network.neutron [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 588.943163] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.053732] env[61970]: DEBUG nova.policy [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '78b9044c170542c58a9a059ffd955f0b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '47196d0ba88c410fb447b74e78c85c1f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 589.288298] env[61970]: DEBUG nova.compute.manager [req-4e135dde-4bb2-4896-96c4-3ac71f112c7e req-66d302a0-3a54-4c0e-b942-70be9a2c2b51 service nova] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Received event network-changed-9ca8ea05-ac30-4e52-8de3-5ec6b39bdc8f {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 589.288298] env[61970]: DEBUG nova.compute.manager [req-4e135dde-4bb2-4896-96c4-3ac71f112c7e req-66d302a0-3a54-4c0e-b942-70be9a2c2b51 service nova] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Refreshing instance network info cache due to event network-changed-9ca8ea05-ac30-4e52-8de3-5ec6b39bdc8f. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 589.288298] env[61970]: DEBUG oslo_concurrency.lockutils [req-4e135dde-4bb2-4896-96c4-3ac71f112c7e req-66d302a0-3a54-4c0e-b942-70be9a2c2b51 service nova] Acquiring lock "refresh_cache-fb0aec23-e010-4f9c-8a1d-1b31d17190b8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.288298] env[61970]: DEBUG oslo_concurrency.lockutils [req-4e135dde-4bb2-4896-96c4-3ac71f112c7e req-66d302a0-3a54-4c0e-b942-70be9a2c2b51 service nova] Acquired lock "refresh_cache-fb0aec23-e010-4f9c-8a1d-1b31d17190b8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.288298] env[61970]: DEBUG nova.network.neutron [req-4e135dde-4bb2-4896-96c4-3ac71f112c7e req-66d302a0-3a54-4c0e-b942-70be9a2c2b51 service nova] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Refreshing network info cache for port 9ca8ea05-ac30-4e52-8de3-5ec6b39bdc8f {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 589.303754] env[61970]: ERROR nova.compute.manager [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9ca8ea05-ac30-4e52-8de3-5ec6b39bdc8f, please check neutron logs for more information. [ 589.303754] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 589.303754] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.303754] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 589.303754] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.303754] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 589.303754] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.303754] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 589.303754] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.303754] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 589.303754] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.303754] env[61970]: ERROR nova.compute.manager raise self.value [ 589.303754] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.303754] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 589.303754] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.303754] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 589.304328] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.304328] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 589.304328] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9ca8ea05-ac30-4e52-8de3-5ec6b39bdc8f, please check neutron logs for more information. [ 589.304328] env[61970]: ERROR nova.compute.manager [ 589.304328] env[61970]: Traceback (most recent call last): [ 589.304328] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 589.304328] env[61970]: listener.cb(fileno) [ 589.304328] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.304328] env[61970]: result = function(*args, **kwargs) [ 589.304328] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.304328] env[61970]: return func(*args, **kwargs) [ 589.304328] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.304328] env[61970]: raise e [ 589.304328] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.304328] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 589.304328] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.304328] env[61970]: created_port_ids = self._update_ports_for_instance( [ 589.304935] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.304935] env[61970]: with excutils.save_and_reraise_exception(): [ 589.304935] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.304935] env[61970]: self.force_reraise() [ 589.304935] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.304935] env[61970]: raise self.value [ 589.304935] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.304935] env[61970]: updated_port = self._update_port( [ 589.304935] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.304935] env[61970]: _ensure_no_port_binding_failure(port) [ 589.304935] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.304935] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 589.304935] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 9ca8ea05-ac30-4e52-8de3-5ec6b39bdc8f, please check neutron logs for more information. [ 589.304935] env[61970]: Removing descriptor: 15 [ 589.306236] env[61970]: ERROR nova.compute.manager [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9ca8ea05-ac30-4e52-8de3-5ec6b39bdc8f, please check neutron logs for more information. [ 589.306236] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Traceback (most recent call last): [ 589.306236] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 589.306236] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] yield resources [ 589.306236] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 589.306236] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] self.driver.spawn(context, instance, image_meta, [ 589.306236] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 589.306236] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.306236] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.306236] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] vm_ref = self.build_virtual_machine(instance, [ 589.306236] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.306620] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.306620] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.306620] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] for vif in network_info: [ 589.306620] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 589.306620] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] return self._sync_wrapper(fn, *args, **kwargs) [ 589.306620] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 589.306620] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] self.wait() [ 589.306620] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 589.306620] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] self[:] = self._gt.wait() [ 589.306620] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.306620] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] return self._exit_event.wait() [ 589.306620] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 589.306620] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] result = hub.switch() [ 589.306990] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 589.306990] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] return self.greenlet.switch() [ 589.306990] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.306990] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] result = function(*args, **kwargs) [ 589.306990] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.306990] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] return func(*args, **kwargs) [ 589.306990] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.306990] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] raise e [ 589.306990] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.306990] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] nwinfo = self.network_api.allocate_for_instance( [ 589.306990] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.306990] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] created_port_ids = self._update_ports_for_instance( [ 589.306990] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.307381] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] with excutils.save_and_reraise_exception(): [ 589.307381] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.307381] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] self.force_reraise() [ 589.307381] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.307381] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] raise self.value [ 589.307381] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.307381] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] updated_port = self._update_port( [ 589.307381] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.307381] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] _ensure_no_port_binding_failure(port) [ 589.307381] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.307381] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] raise exception.PortBindingFailed(port_id=port['id']) [ 589.307381] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] nova.exception.PortBindingFailed: Binding failed for port 9ca8ea05-ac30-4e52-8de3-5ec6b39bdc8f, please check neutron logs for more information. [ 589.307381] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] [ 589.307779] env[61970]: INFO nova.compute.manager [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Terminating instance [ 589.309881] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Acquiring lock "refresh_cache-fb0aec23-e010-4f9c-8a1d-1b31d17190b8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.345634] env[61970]: DEBUG nova.compute.manager [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 589.765223] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d552e86a-8803-4f53-87a2-dfef93de5520 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.776437] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-406a9859-93f6-4267-aa10-a094ca153fa0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.815975] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8f3a3ae-b4b0-4d7e-91cb-449ee716b65f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.824087] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b3f811-8638-4658-b3f4-69c4a2f4a825 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.845704] env[61970]: DEBUG nova.compute.provider_tree [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 589.858623] env[61970]: DEBUG oslo_concurrency.lockutils [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "719de3af-ffe0-4352-9d2c-9e102859fb6c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.858865] env[61970]: DEBUG oslo_concurrency.lockutils [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "719de3af-ffe0-4352-9d2c-9e102859fb6c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.893088] env[61970]: DEBUG nova.network.neutron [req-4e135dde-4bb2-4896-96c4-3ac71f112c7e req-66d302a0-3a54-4c0e-b942-70be9a2c2b51 service nova] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.161648] env[61970]: DEBUG nova.network.neutron [req-4e135dde-4bb2-4896-96c4-3ac71f112c7e req-66d302a0-3a54-4c0e-b942-70be9a2c2b51 service nova] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.278505] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Acquiring lock "b205db9a-8170-4c53-8212-f3fc2ec11d1c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.278723] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Lock "b205db9a-8170-4c53-8212-f3fc2ec11d1c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.350499] env[61970]: DEBUG nova.scheduler.client.report [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.357811] env[61970]: DEBUG nova.compute.manager [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 590.389474] env[61970]: DEBUG nova.virt.hardware [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 590.390292] env[61970]: DEBUG nova.virt.hardware [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 590.390569] env[61970]: DEBUG nova.virt.hardware [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 590.390853] env[61970]: DEBUG nova.virt.hardware [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 590.391370] env[61970]: DEBUG nova.virt.hardware [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 590.391370] env[61970]: DEBUG nova.virt.hardware [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 590.391370] env[61970]: DEBUG nova.virt.hardware [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 590.391566] env[61970]: DEBUG nova.virt.hardware [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 590.391690] env[61970]: DEBUG nova.virt.hardware [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 590.392123] env[61970]: DEBUG nova.virt.hardware [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 590.392123] env[61970]: DEBUG nova.virt.hardware [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 590.394595] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-245c5d14-22df-4e08-a739-f3d3a5376a39 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.401445] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28137eb2-30f3-4112-98e8-a412883fab68 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.429023] env[61970]: DEBUG nova.network.neutron [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Successfully created port: a4134d35-8c9e-43b0-8491-44c9fc75a801 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 590.665315] env[61970]: DEBUG oslo_concurrency.lockutils [req-4e135dde-4bb2-4896-96c4-3ac71f112c7e req-66d302a0-3a54-4c0e-b942-70be9a2c2b51 service nova] Releasing lock "refresh_cache-fb0aec23-e010-4f9c-8a1d-1b31d17190b8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.665766] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Acquired lock "refresh_cache-fb0aec23-e010-4f9c-8a1d-1b31d17190b8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.665957] env[61970]: DEBUG nova.network.neutron [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 590.855552] env[61970]: DEBUG oslo_concurrency.lockutils [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.533s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.856081] env[61970]: DEBUG nova.compute.manager [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 590.860540] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.747s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.205316] env[61970]: DEBUG nova.network.neutron [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.363897] env[61970]: DEBUG nova.compute.utils [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 591.373598] env[61970]: DEBUG nova.compute.manager [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 591.373598] env[61970]: DEBUG nova.network.neutron [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 591.480934] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Acquiring lock "84c6fdb9-0858-4f25-bdc8-474ffb9f6abb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.482076] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Lock "84c6fdb9-0858-4f25-bdc8-474ffb9f6abb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.490296] env[61970]: DEBUG nova.network.neutron [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.521039] env[61970]: DEBUG nova.policy [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '513b373d7dc74908a71957e05347af8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5640a0fe1fd0404bbe28cbdccbc8311f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 591.776650] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c092ec21-f161-4070-ae7d-555aa6867204 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.787567] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c36b6f5b-41fc-40e7-ad1e-8d1fc3870e82 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.823102] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-595a4082-4eb3-485f-8b4b-2c3afc768cd4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.831486] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f44604-7be9-4712-bd35-fa8418796cde {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.847038] env[61970]: DEBUG nova.compute.provider_tree [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.871563] env[61970]: DEBUG nova.compute.manager [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 591.994029] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Releasing lock "refresh_cache-fb0aec23-e010-4f9c-8a1d-1b31d17190b8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.994289] env[61970]: DEBUG nova.compute.manager [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 591.994518] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 591.995801] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a2603723-0d08-47fd-b88a-a64fd3ed14fd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.006116] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e1c3c3-9969-4d1a-b595-b51fc9e95d87 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.029079] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fb0aec23-e010-4f9c-8a1d-1b31d17190b8 could not be found. [ 592.029689] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 592.029689] env[61970]: INFO nova.compute.manager [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 592.029844] env[61970]: DEBUG oslo.service.loopingcall [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 592.030033] env[61970]: DEBUG nova.compute.manager [-] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 592.030115] env[61970]: DEBUG nova.network.neutron [-] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 592.110521] env[61970]: DEBUG nova.network.neutron [-] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.233905] env[61970]: DEBUG nova.network.neutron [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Successfully created port: a8e611aa-9ddc-45d3-a73a-619ee9d10767 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 592.273833] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Acquiring lock "9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.274065] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Lock "9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.344876] env[61970]: DEBUG nova.compute.manager [req-e842a7e3-f4d2-4079-a35c-d4202d7d70e3 req-0acbcfb1-9166-494e-87d2-8a5ddf0bc39c service nova] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Received event network-vif-deleted-9ca8ea05-ac30-4e52-8de3-5ec6b39bdc8f {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 592.353147] env[61970]: DEBUG nova.scheduler.client.report [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.589958] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Acquiring lock "3661a95f-a75c-4f0f-ad25-428afdb7bccb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.590218] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Lock "3661a95f-a75c-4f0f-ad25-428afdb7bccb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.615081] env[61970]: DEBUG nova.network.neutron [-] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.859863] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.860503] env[61970]: ERROR nova.compute.manager [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9886c8df-135a-4a75-bac1-b546e16ef653, please check neutron logs for more information. [ 592.860503] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Traceback (most recent call last): [ 592.860503] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 592.860503] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] self.driver.spawn(context, instance, image_meta, [ 592.860503] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 592.860503] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.860503] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.860503] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] vm_ref = self.build_virtual_machine(instance, [ 592.860503] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.860503] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.860503] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.860923] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] for vif in network_info: [ 592.860923] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 592.860923] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] return self._sync_wrapper(fn, *args, **kwargs) [ 592.860923] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 592.860923] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] self.wait() [ 592.860923] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 592.860923] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] self[:] = self._gt.wait() [ 592.860923] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.860923] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] return self._exit_event.wait() [ 592.860923] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 592.860923] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] result = hub.switch() [ 592.860923] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 592.860923] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] return self.greenlet.switch() [ 592.861531] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.861531] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] result = function(*args, **kwargs) [ 592.861531] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 592.861531] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] return func(*args, **kwargs) [ 592.861531] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.861531] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] raise e [ 592.861531] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.861531] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] nwinfo = self.network_api.allocate_for_instance( [ 592.861531] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.861531] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] created_port_ids = self._update_ports_for_instance( [ 592.861531] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.861531] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] with excutils.save_and_reraise_exception(): [ 592.861531] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.861961] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] self.force_reraise() [ 592.861961] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.861961] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] raise self.value [ 592.861961] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.861961] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] updated_port = self._update_port( [ 592.861961] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.861961] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] _ensure_no_port_binding_failure(port) [ 592.861961] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.861961] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] raise exception.PortBindingFailed(port_id=port['id']) [ 592.861961] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] nova.exception.PortBindingFailed: Binding failed for port 9886c8df-135a-4a75-bac1-b546e16ef653, please check neutron logs for more information. [ 592.861961] env[61970]: ERROR nova.compute.manager [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] [ 592.862826] env[61970]: DEBUG nova.compute.utils [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Binding failed for port 9886c8df-135a-4a75-bac1-b546e16ef653, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 592.864679] env[61970]: DEBUG nova.compute.manager [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Build of instance 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f was re-scheduled: Binding failed for port 9886c8df-135a-4a75-bac1-b546e16ef653, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 592.866008] env[61970]: DEBUG nova.compute.manager [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 592.867123] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Acquiring lock "refresh_cache-604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.867300] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Acquired lock "refresh_cache-604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.867476] env[61970]: DEBUG nova.network.neutron [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 592.868607] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.379s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.872459] env[61970]: INFO nova.compute.claims [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 592.881204] env[61970]: DEBUG nova.compute.manager [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 592.913701] env[61970]: DEBUG nova.virt.hardware [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 592.914065] env[61970]: DEBUG nova.virt.hardware [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 592.914065] env[61970]: DEBUG nova.virt.hardware [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 592.914251] env[61970]: DEBUG nova.virt.hardware [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 592.914391] env[61970]: DEBUG nova.virt.hardware [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 592.914528] env[61970]: DEBUG nova.virt.hardware [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 592.914767] env[61970]: DEBUG nova.virt.hardware [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 592.914876] env[61970]: DEBUG nova.virt.hardware [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 592.915305] env[61970]: DEBUG nova.virt.hardware [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 592.915305] env[61970]: DEBUG nova.virt.hardware [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 592.915421] env[61970]: DEBUG nova.virt.hardware [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 592.916678] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd5f705-303a-4f74-bf08-3af6fa5a8b08 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.925743] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98678410-897f-4e41-8196-305c34aea418 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.117421] env[61970]: INFO nova.compute.manager [-] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Took 1.09 seconds to deallocate network for instance. [ 593.128503] env[61970]: DEBUG nova.compute.claims [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 593.128785] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.420805] env[61970]: DEBUG nova.network.neutron [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.732761] env[61970]: DEBUG nova.network.neutron [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.743440] env[61970]: ERROR nova.compute.manager [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a4134d35-8c9e-43b0-8491-44c9fc75a801, please check neutron logs for more information. [ 593.743440] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 593.743440] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.743440] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 593.743440] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.743440] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 593.743440] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.743440] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 593.743440] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.743440] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 593.743440] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.743440] env[61970]: ERROR nova.compute.manager raise self.value [ 593.743440] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.743440] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 593.743440] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.743440] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 593.746454] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.746454] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 593.746454] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a4134d35-8c9e-43b0-8491-44c9fc75a801, please check neutron logs for more information. [ 593.746454] env[61970]: ERROR nova.compute.manager [ 593.746454] env[61970]: Traceback (most recent call last): [ 593.746454] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 593.746454] env[61970]: listener.cb(fileno) [ 593.746454] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.746454] env[61970]: result = function(*args, **kwargs) [ 593.746454] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 593.746454] env[61970]: return func(*args, **kwargs) [ 593.746454] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.746454] env[61970]: raise e [ 593.746454] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.746454] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 593.746454] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.746454] env[61970]: created_port_ids = self._update_ports_for_instance( [ 593.746454] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.746454] env[61970]: with excutils.save_and_reraise_exception(): [ 593.746454] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.746454] env[61970]: self.force_reraise() [ 593.746454] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.746454] env[61970]: raise self.value [ 593.746454] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.746454] env[61970]: updated_port = self._update_port( [ 593.746454] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.746454] env[61970]: _ensure_no_port_binding_failure(port) [ 593.746454] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.746454] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 593.748383] env[61970]: nova.exception.PortBindingFailed: Binding failed for port a4134d35-8c9e-43b0-8491-44c9fc75a801, please check neutron logs for more information. [ 593.748383] env[61970]: Removing descriptor: 17 [ 593.748383] env[61970]: ERROR nova.compute.manager [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a4134d35-8c9e-43b0-8491-44c9fc75a801, please check neutron logs for more information. [ 593.748383] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Traceback (most recent call last): [ 593.748383] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 593.748383] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] yield resources [ 593.748383] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 593.748383] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] self.driver.spawn(context, instance, image_meta, [ 593.748383] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 593.748383] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 593.748383] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 593.748383] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] vm_ref = self.build_virtual_machine(instance, [ 593.749167] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 593.749167] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] vif_infos = vmwarevif.get_vif_info(self._session, [ 593.749167] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 593.749167] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] for vif in network_info: [ 593.749167] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 593.749167] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] return self._sync_wrapper(fn, *args, **kwargs) [ 593.749167] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 593.749167] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] self.wait() [ 593.749167] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 593.749167] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] self[:] = self._gt.wait() [ 593.749167] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 593.749167] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] return self._exit_event.wait() [ 593.749167] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 593.749551] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] result = hub.switch() [ 593.749551] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 593.749551] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] return self.greenlet.switch() [ 593.749551] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.749551] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] result = function(*args, **kwargs) [ 593.749551] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 593.749551] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] return func(*args, **kwargs) [ 593.749551] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.749551] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] raise e [ 593.749551] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.749551] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] nwinfo = self.network_api.allocate_for_instance( [ 593.749551] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.749551] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] created_port_ids = self._update_ports_for_instance( [ 593.749926] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.749926] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] with excutils.save_and_reraise_exception(): [ 593.749926] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.749926] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] self.force_reraise() [ 593.749926] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.749926] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] raise self.value [ 593.749926] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.749926] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] updated_port = self._update_port( [ 593.749926] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.749926] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] _ensure_no_port_binding_failure(port) [ 593.749926] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.749926] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] raise exception.PortBindingFailed(port_id=port['id']) [ 593.751367] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] nova.exception.PortBindingFailed: Binding failed for port a4134d35-8c9e-43b0-8491-44c9fc75a801, please check neutron logs for more information. [ 593.751367] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] [ 593.751367] env[61970]: INFO nova.compute.manager [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Terminating instance [ 593.752208] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Acquiring lock "refresh_cache-de9ed575-9e43-4c67-89ed-0549b43fb81a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.752379] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Acquired lock "refresh_cache-de9ed575-9e43-4c67-89ed-0549b43fb81a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.752551] env[61970]: DEBUG nova.network.neutron [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 593.774165] env[61970]: ERROR nova.compute.manager [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a8e611aa-9ddc-45d3-a73a-619ee9d10767, please check neutron logs for more information. [ 593.774165] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 593.774165] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.774165] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 593.774165] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.774165] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 593.774165] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.774165] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 593.774165] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.774165] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 593.774165] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.774165] env[61970]: ERROR nova.compute.manager raise self.value [ 593.774165] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.774165] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 593.774165] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.774165] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 593.774890] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.774890] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 593.774890] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a8e611aa-9ddc-45d3-a73a-619ee9d10767, please check neutron logs for more information. [ 593.774890] env[61970]: ERROR nova.compute.manager [ 593.774890] env[61970]: Traceback (most recent call last): [ 593.774890] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 593.774890] env[61970]: listener.cb(fileno) [ 593.774890] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.774890] env[61970]: result = function(*args, **kwargs) [ 593.774890] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 593.774890] env[61970]: return func(*args, **kwargs) [ 593.774890] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.774890] env[61970]: raise e [ 593.774890] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.774890] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 593.774890] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.774890] env[61970]: created_port_ids = self._update_ports_for_instance( [ 593.774890] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.774890] env[61970]: with excutils.save_and_reraise_exception(): [ 593.774890] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.774890] env[61970]: self.force_reraise() [ 593.774890] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.774890] env[61970]: raise self.value [ 593.774890] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.774890] env[61970]: updated_port = self._update_port( [ 593.774890] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.774890] env[61970]: _ensure_no_port_binding_failure(port) [ 593.774890] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.774890] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 593.776255] env[61970]: nova.exception.PortBindingFailed: Binding failed for port a8e611aa-9ddc-45d3-a73a-619ee9d10767, please check neutron logs for more information. [ 593.776255] env[61970]: Removing descriptor: 20 [ 593.776255] env[61970]: ERROR nova.compute.manager [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a8e611aa-9ddc-45d3-a73a-619ee9d10767, please check neutron logs for more information. [ 593.776255] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Traceback (most recent call last): [ 593.776255] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 593.776255] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] yield resources [ 593.776255] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 593.776255] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] self.driver.spawn(context, instance, image_meta, [ 593.776255] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 593.776255] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 593.776255] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 593.776255] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] vm_ref = self.build_virtual_machine(instance, [ 593.777046] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 593.777046] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] vif_infos = vmwarevif.get_vif_info(self._session, [ 593.777046] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 593.777046] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] for vif in network_info: [ 593.777046] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 593.777046] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] return self._sync_wrapper(fn, *args, **kwargs) [ 593.777046] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 593.777046] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] self.wait() [ 593.777046] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 593.777046] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] self[:] = self._gt.wait() [ 593.777046] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 593.777046] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] return self._exit_event.wait() [ 593.777046] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 593.777711] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] result = hub.switch() [ 593.777711] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 593.777711] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] return self.greenlet.switch() [ 593.777711] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.777711] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] result = function(*args, **kwargs) [ 593.777711] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 593.777711] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] return func(*args, **kwargs) [ 593.777711] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.777711] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] raise e [ 593.777711] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.777711] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] nwinfo = self.network_api.allocate_for_instance( [ 593.777711] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.777711] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] created_port_ids = self._update_ports_for_instance( [ 593.778294] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.778294] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] with excutils.save_and_reraise_exception(): [ 593.778294] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.778294] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] self.force_reraise() [ 593.778294] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.778294] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] raise self.value [ 593.778294] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.778294] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] updated_port = self._update_port( [ 593.778294] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.778294] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] _ensure_no_port_binding_failure(port) [ 593.778294] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.778294] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] raise exception.PortBindingFailed(port_id=port['id']) [ 593.778740] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] nova.exception.PortBindingFailed: Binding failed for port a8e611aa-9ddc-45d3-a73a-619ee9d10767, please check neutron logs for more information. [ 593.778740] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] [ 593.778740] env[61970]: INFO nova.compute.manager [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Terminating instance [ 593.778740] env[61970]: DEBUG oslo_concurrency.lockutils [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Acquiring lock "refresh_cache-0ef75e57-83af-4440-890e-95dcdb9920d2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.778740] env[61970]: DEBUG oslo_concurrency.lockutils [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Acquired lock "refresh_cache-0ef75e57-83af-4440-890e-95dcdb9920d2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.778740] env[61970]: DEBUG nova.network.neutron [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 594.054920] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Acquiring lock "7dda08a4-83ae-4882-9469-324d948c9749" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.055158] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Lock "7dda08a4-83ae-4882-9469-324d948c9749" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.140658] env[61970]: DEBUG nova.compute.manager [req-0d1b88ab-97fc-4f6d-8d9f-33961e81add8 req-603227cf-1acf-4ac3-ad6a-6293df41c01e service nova] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Received event network-changed-a8e611aa-9ddc-45d3-a73a-619ee9d10767 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 594.140658] env[61970]: DEBUG nova.compute.manager [req-0d1b88ab-97fc-4f6d-8d9f-33961e81add8 req-603227cf-1acf-4ac3-ad6a-6293df41c01e service nova] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Refreshing instance network info cache due to event network-changed-a8e611aa-9ddc-45d3-a73a-619ee9d10767. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 594.140658] env[61970]: DEBUG oslo_concurrency.lockutils [req-0d1b88ab-97fc-4f6d-8d9f-33961e81add8 req-603227cf-1acf-4ac3-ad6a-6293df41c01e service nova] Acquiring lock "refresh_cache-0ef75e57-83af-4440-890e-95dcdb9920d2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.239273] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Releasing lock "refresh_cache-604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.239273] env[61970]: DEBUG nova.compute.manager [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 594.239273] env[61970]: DEBUG nova.compute.manager [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 594.239273] env[61970]: DEBUG nova.network.neutron [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 594.288086] env[61970]: DEBUG nova.network.neutron [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.291242] env[61970]: DEBUG nova.network.neutron [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.334908] env[61970]: DEBUG nova.network.neutron [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.386478] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-046f2602-24a0-4c21-9d9b-2c060409fb16 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.394532] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eefe9d1c-737e-47cf-b01d-8621451eb793 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.430554] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbfd59bd-2ace-4ccf-b2ec-8dabf8e2f497 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.439969] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05204086-62a1-4ea9-8017-7df99f62b759 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.453782] env[61970]: DEBUG nova.compute.provider_tree [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.462902] env[61970]: DEBUG nova.network.neutron [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.501240] env[61970]: DEBUG nova.compute.manager [req-cd3beeb9-e742-4926-b1cc-cd0068fc53eb req-59de026a-3b0a-4780-9287-768baca26eab service nova] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Received event network-changed-a4134d35-8c9e-43b0-8491-44c9fc75a801 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 594.501579] env[61970]: DEBUG nova.compute.manager [req-cd3beeb9-e742-4926-b1cc-cd0068fc53eb req-59de026a-3b0a-4780-9287-768baca26eab service nova] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Refreshing instance network info cache due to event network-changed-a4134d35-8c9e-43b0-8491-44c9fc75a801. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 594.502137] env[61970]: DEBUG oslo_concurrency.lockutils [req-cd3beeb9-e742-4926-b1cc-cd0068fc53eb req-59de026a-3b0a-4780-9287-768baca26eab service nova] Acquiring lock "refresh_cache-de9ed575-9e43-4c67-89ed-0549b43fb81a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.507327] env[61970]: DEBUG nova.network.neutron [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.799552] env[61970]: DEBUG nova.network.neutron [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.957473] env[61970]: DEBUG nova.scheduler.client.report [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.968628] env[61970]: DEBUG oslo_concurrency.lockutils [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Releasing lock "refresh_cache-0ef75e57-83af-4440-890e-95dcdb9920d2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.969094] env[61970]: DEBUG nova.compute.manager [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 594.969314] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 594.970468] env[61970]: DEBUG oslo_concurrency.lockutils [req-0d1b88ab-97fc-4f6d-8d9f-33961e81add8 req-603227cf-1acf-4ac3-ad6a-6293df41c01e service nova] Acquired lock "refresh_cache-0ef75e57-83af-4440-890e-95dcdb9920d2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.970715] env[61970]: DEBUG nova.network.neutron [req-0d1b88ab-97fc-4f6d-8d9f-33961e81add8 req-603227cf-1acf-4ac3-ad6a-6293df41c01e service nova] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Refreshing network info cache for port a8e611aa-9ddc-45d3-a73a-619ee9d10767 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 594.972131] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-85d43e96-aa26-4292-a900-637acce16a04 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.981915] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a47caa4-c69e-4997-b130-31e67abb6b49 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.008015] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0ef75e57-83af-4440-890e-95dcdb9920d2 could not be found. [ 595.008015] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 595.008015] env[61970]: INFO nova.compute.manager [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 595.008222] env[61970]: DEBUG oslo.service.loopingcall [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 595.008361] env[61970]: DEBUG nova.compute.manager [-] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 595.008812] env[61970]: DEBUG nova.network.neutron [-] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 595.010810] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Releasing lock "refresh_cache-de9ed575-9e43-4c67-89ed-0549b43fb81a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.010810] env[61970]: DEBUG nova.compute.manager [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 595.010936] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 595.011195] env[61970]: DEBUG oslo_concurrency.lockutils [req-cd3beeb9-e742-4926-b1cc-cd0068fc53eb req-59de026a-3b0a-4780-9287-768baca26eab service nova] Acquired lock "refresh_cache-de9ed575-9e43-4c67-89ed-0549b43fb81a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.011358] env[61970]: DEBUG nova.network.neutron [req-cd3beeb9-e742-4926-b1cc-cd0068fc53eb req-59de026a-3b0a-4780-9287-768baca26eab service nova] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Refreshing network info cache for port a4134d35-8c9e-43b0-8491-44c9fc75a801 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 595.012353] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a4cddbdf-e6b8-41f4-bb22-2551cda40dc7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.020765] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b888df7a-2c9c-4fec-a3f9-1e6c9b0e272e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.031613] env[61970]: DEBUG nova.network.neutron [-] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.046028] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance de9ed575-9e43-4c67-89ed-0549b43fb81a could not be found. [ 595.046256] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 595.046438] env[61970]: INFO nova.compute.manager [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 595.046671] env[61970]: DEBUG oslo.service.loopingcall [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 595.047139] env[61970]: DEBUG nova.compute.manager [-] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 595.047236] env[61970]: DEBUG nova.network.neutron [-] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 595.086674] env[61970]: DEBUG nova.network.neutron [-] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.304217] env[61970]: INFO nova.compute.manager [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] [instance: 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f] Took 1.06 seconds to deallocate network for instance. [ 595.463324] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.594s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.463797] env[61970]: DEBUG nova.compute.manager [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 595.467807] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.565s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.535610] env[61970]: DEBUG nova.network.neutron [-] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.589693] env[61970]: DEBUG nova.network.neutron [-] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.680830] env[61970]: DEBUG nova.network.neutron [req-0d1b88ab-97fc-4f6d-8d9f-33961e81add8 req-603227cf-1acf-4ac3-ad6a-6293df41c01e service nova] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.683473] env[61970]: DEBUG nova.network.neutron [req-cd3beeb9-e742-4926-b1cc-cd0068fc53eb req-59de026a-3b0a-4780-9287-768baca26eab service nova] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.789785] env[61970]: DEBUG nova.network.neutron [req-0d1b88ab-97fc-4f6d-8d9f-33961e81add8 req-603227cf-1acf-4ac3-ad6a-6293df41c01e service nova] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.858517] env[61970]: DEBUG nova.network.neutron [req-cd3beeb9-e742-4926-b1cc-cd0068fc53eb req-59de026a-3b0a-4780-9287-768baca26eab service nova] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.975529] env[61970]: DEBUG nova.compute.utils [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 595.977367] env[61970]: DEBUG nova.compute.manager [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 595.977620] env[61970]: DEBUG nova.network.neutron [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 596.040436] env[61970]: INFO nova.compute.manager [-] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Took 1.03 seconds to deallocate network for instance. [ 596.042942] env[61970]: DEBUG nova.compute.claims [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 596.043133] env[61970]: DEBUG oslo_concurrency.lockutils [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.080029] env[61970]: DEBUG nova.policy [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '77ff100dcaf845648b77586a2205c9ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c642e7570e8248d3aae5fb739e9a928d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 596.094823] env[61970]: INFO nova.compute.manager [-] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Took 1.05 seconds to deallocate network for instance. [ 596.098354] env[61970]: DEBUG nova.compute.claims [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 596.098453] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.292014] env[61970]: DEBUG oslo_concurrency.lockutils [req-0d1b88ab-97fc-4f6d-8d9f-33961e81add8 req-603227cf-1acf-4ac3-ad6a-6293df41c01e service nova] Releasing lock "refresh_cache-0ef75e57-83af-4440-890e-95dcdb9920d2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.330206] env[61970]: DEBUG nova.compute.manager [req-5dbea711-d125-49d4-8dd3-ef48e5b8c076 req-ed51833a-5ce0-451a-9ab2-bb0e6f6e54ad service nova] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Received event network-vif-deleted-a8e611aa-9ddc-45d3-a73a-619ee9d10767 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 596.352627] env[61970]: INFO nova.scheduler.client.report [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Deleted allocations for instance 604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f [ 596.359857] env[61970]: DEBUG oslo_concurrency.lockutils [req-cd3beeb9-e742-4926-b1cc-cd0068fc53eb req-59de026a-3b0a-4780-9287-768baca26eab service nova] Releasing lock "refresh_cache-de9ed575-9e43-4c67-89ed-0549b43fb81a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.360173] env[61970]: DEBUG nova.compute.manager [req-cd3beeb9-e742-4926-b1cc-cd0068fc53eb req-59de026a-3b0a-4780-9287-768baca26eab service nova] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Received event network-vif-deleted-a4134d35-8c9e-43b0-8491-44c9fc75a801 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 596.465585] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e392c719-c45a-476d-81d9-5a181b39d721 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.475278] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-784663ae-9bd8-4492-9fcb-19c076374b8e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.481816] env[61970]: DEBUG nova.compute.manager [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 596.521855] env[61970]: DEBUG nova.network.neutron [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Successfully created port: fd66c242-f49a-4324-ac18-fba690484aa7 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 596.524548] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c948fa3a-9f17-4c15-9254-6088daadc952 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.534296] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25effb41-dc84-4e57-afe5-70055ccebae7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.550032] env[61970]: DEBUG nova.compute.provider_tree [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.865187] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9dc08285-0026-4b74-a7b8-bcf4dca0706e tempest-ServersWithSpecificFlavorTestJSON-880293145 tempest-ServersWithSpecificFlavorTestJSON-880293145-project-member] Lock "604d33e8-58ab-4c8a-be5f-b7cfe6a9fe4f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.620s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.054518] env[61970]: DEBUG nova.scheduler.client.report [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.370600] env[61970]: DEBUG nova.compute.manager [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 597.528546] env[61970]: DEBUG nova.compute.manager [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 597.561249] env[61970]: DEBUG nova.virt.hardware [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 597.561556] env[61970]: DEBUG nova.virt.hardware [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 597.561743] env[61970]: DEBUG nova.virt.hardware [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.561953] env[61970]: DEBUG nova.virt.hardware [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 597.562134] env[61970]: DEBUG nova.virt.hardware [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.562310] env[61970]: DEBUG nova.virt.hardware [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 597.562627] env[61970]: DEBUG nova.virt.hardware [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 597.562834] env[61970]: DEBUG nova.virt.hardware [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 597.563040] env[61970]: DEBUG nova.virt.hardware [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 597.563233] env[61970]: DEBUG nova.virt.hardware [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 597.563468] env[61970]: DEBUG nova.virt.hardware [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 597.564222] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.097s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.564863] env[61970]: ERROR nova.compute.manager [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 69991011-5146-47f0-a8a2-794791aa05ec, please check neutron logs for more information. [ 597.564863] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Traceback (most recent call last): [ 597.564863] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.564863] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] self.driver.spawn(context, instance, image_meta, [ 597.564863] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 597.564863] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.564863] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.564863] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] vm_ref = self.build_virtual_machine(instance, [ 597.564863] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.564863] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.564863] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.565238] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] for vif in network_info: [ 597.565238] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 597.565238] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] return self._sync_wrapper(fn, *args, **kwargs) [ 597.565238] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 597.565238] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] self.wait() [ 597.565238] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 597.565238] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] self[:] = self._gt.wait() [ 597.565238] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.565238] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] return self._exit_event.wait() [ 597.565238] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.565238] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] result = hub.switch() [ 597.565238] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.565238] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] return self.greenlet.switch() [ 597.565642] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.565642] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] result = function(*args, **kwargs) [ 597.565642] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 597.565642] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] return func(*args, **kwargs) [ 597.565642] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.565642] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] raise e [ 597.565642] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.565642] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] nwinfo = self.network_api.allocate_for_instance( [ 597.565642] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.565642] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] created_port_ids = self._update_ports_for_instance( [ 597.565642] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.565642] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] with excutils.save_and_reraise_exception(): [ 597.565642] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.565994] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] self.force_reraise() [ 597.565994] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.565994] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] raise self.value [ 597.565994] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.565994] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] updated_port = self._update_port( [ 597.565994] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.565994] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] _ensure_no_port_binding_failure(port) [ 597.565994] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.565994] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] raise exception.PortBindingFailed(port_id=port['id']) [ 597.565994] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] nova.exception.PortBindingFailed: Binding failed for port 69991011-5146-47f0-a8a2-794791aa05ec, please check neutron logs for more information. [ 597.565994] env[61970]: ERROR nova.compute.manager [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] [ 597.566316] env[61970]: DEBUG nova.compute.utils [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Binding failed for port 69991011-5146-47f0-a8a2-794791aa05ec, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 597.567426] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e991fb9-9f4b-4730-97fe-f6416dc391df {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.570863] env[61970]: DEBUG nova.compute.manager [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Build of instance 39bc8558-ce7f-41e9-b913-9b628a220ca7 was re-scheduled: Binding failed for port 69991011-5146-47f0-a8a2-794791aa05ec, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 597.571364] env[61970]: DEBUG nova.compute.manager [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 597.571630] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Acquiring lock "refresh_cache-39bc8558-ce7f-41e9-b913-9b628a220ca7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.571802] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Acquired lock "refresh_cache-39bc8558-ce7f-41e9-b913-9b628a220ca7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.571982] env[61970]: DEBUG nova.network.neutron [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 597.580205] env[61970]: DEBUG oslo_concurrency.lockutils [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.871s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.580205] env[61970]: INFO nova.compute.claims [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 597.591559] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-595a4ae2-2f26-4421-8a65-0e4414720c59 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.823764] env[61970]: ERROR nova.compute.manager [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fd66c242-f49a-4324-ac18-fba690484aa7, please check neutron logs for more information. [ 597.823764] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 597.823764] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.823764] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 597.823764] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.823764] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 597.823764] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.823764] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 597.823764] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.823764] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 597.823764] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.823764] env[61970]: ERROR nova.compute.manager raise self.value [ 597.823764] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.823764] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 597.823764] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.823764] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 597.824368] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.824368] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 597.824368] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fd66c242-f49a-4324-ac18-fba690484aa7, please check neutron logs for more information. [ 597.824368] env[61970]: ERROR nova.compute.manager [ 597.824368] env[61970]: Traceback (most recent call last): [ 597.824368] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 597.824368] env[61970]: listener.cb(fileno) [ 597.824368] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.824368] env[61970]: result = function(*args, **kwargs) [ 597.824368] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 597.824368] env[61970]: return func(*args, **kwargs) [ 597.824368] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.824368] env[61970]: raise e [ 597.824368] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.824368] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 597.824368] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.824368] env[61970]: created_port_ids = self._update_ports_for_instance( [ 597.824368] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.824368] env[61970]: with excutils.save_and_reraise_exception(): [ 597.824368] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.824368] env[61970]: self.force_reraise() [ 597.824368] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.824368] env[61970]: raise self.value [ 597.824368] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.824368] env[61970]: updated_port = self._update_port( [ 597.824368] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.824368] env[61970]: _ensure_no_port_binding_failure(port) [ 597.824368] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.824368] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 597.825326] env[61970]: nova.exception.PortBindingFailed: Binding failed for port fd66c242-f49a-4324-ac18-fba690484aa7, please check neutron logs for more information. [ 597.825326] env[61970]: Removing descriptor: 15 [ 597.825326] env[61970]: ERROR nova.compute.manager [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fd66c242-f49a-4324-ac18-fba690484aa7, please check neutron logs for more information. [ 597.825326] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Traceback (most recent call last): [ 597.825326] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 597.825326] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] yield resources [ 597.825326] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.825326] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] self.driver.spawn(context, instance, image_meta, [ 597.825326] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 597.825326] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.825326] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.825326] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] vm_ref = self.build_virtual_machine(instance, [ 597.825777] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.825777] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.825777] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.825777] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] for vif in network_info: [ 597.825777] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 597.825777] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] return self._sync_wrapper(fn, *args, **kwargs) [ 597.825777] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 597.825777] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] self.wait() [ 597.825777] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 597.825777] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] self[:] = self._gt.wait() [ 597.825777] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.825777] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] return self._exit_event.wait() [ 597.825777] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.826230] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] result = hub.switch() [ 597.826230] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.826230] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] return self.greenlet.switch() [ 597.826230] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.826230] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] result = function(*args, **kwargs) [ 597.826230] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 597.826230] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] return func(*args, **kwargs) [ 597.826230] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.826230] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] raise e [ 597.826230] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.826230] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] nwinfo = self.network_api.allocate_for_instance( [ 597.826230] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.826230] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] created_port_ids = self._update_ports_for_instance( [ 597.826691] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.826691] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] with excutils.save_and_reraise_exception(): [ 597.826691] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.826691] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] self.force_reraise() [ 597.826691] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.826691] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] raise self.value [ 597.826691] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.826691] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] updated_port = self._update_port( [ 597.826691] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.826691] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] _ensure_no_port_binding_failure(port) [ 597.826691] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.826691] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] raise exception.PortBindingFailed(port_id=port['id']) [ 597.827111] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] nova.exception.PortBindingFailed: Binding failed for port fd66c242-f49a-4324-ac18-fba690484aa7, please check neutron logs for more information. [ 597.827111] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] [ 597.827111] env[61970]: INFO nova.compute.manager [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Terminating instance [ 597.827111] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Acquiring lock "refresh_cache-d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.827111] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Acquired lock "refresh_cache-d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.827290] env[61970]: DEBUG nova.network.neutron [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 597.857949] env[61970]: DEBUG nova.compute.manager [req-ec553b84-d0d0-4cf8-bb49-4df3509ab20f req-23227732-7e32-4f02-abb2-6221707a3b1f service nova] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Received event network-changed-fd66c242-f49a-4324-ac18-fba690484aa7 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 597.858155] env[61970]: DEBUG nova.compute.manager [req-ec553b84-d0d0-4cf8-bb49-4df3509ab20f req-23227732-7e32-4f02-abb2-6221707a3b1f service nova] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Refreshing instance network info cache due to event network-changed-fd66c242-f49a-4324-ac18-fba690484aa7. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 597.858336] env[61970]: DEBUG oslo_concurrency.lockutils [req-ec553b84-d0d0-4cf8-bb49-4df3509ab20f req-23227732-7e32-4f02-abb2-6221707a3b1f service nova] Acquiring lock "refresh_cache-d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.892956] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.118573] env[61970]: DEBUG nova.network.neutron [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.247883] env[61970]: DEBUG oslo_concurrency.lockutils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Acquiring lock "9d71be92-36ef-4523-919a-931110d73248" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.248603] env[61970]: DEBUG oslo_concurrency.lockutils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Lock "9d71be92-36ef-4523-919a-931110d73248" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.291919] env[61970]: DEBUG nova.network.neutron [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.358703] env[61970]: DEBUG nova.network.neutron [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.449354] env[61970]: DEBUG nova.network.neutron [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.793839] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Releasing lock "refresh_cache-39bc8558-ce7f-41e9-b913-9b628a220ca7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.794132] env[61970]: DEBUG nova.compute.manager [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 598.794267] env[61970]: DEBUG nova.compute.manager [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.794415] env[61970]: DEBUG nova.network.neutron [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 598.811129] env[61970]: DEBUG nova.network.neutron [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.952092] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Releasing lock "refresh_cache-d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.952554] env[61970]: DEBUG nova.compute.manager [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 598.952681] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 598.953919] env[61970]: DEBUG oslo_concurrency.lockutils [req-ec553b84-d0d0-4cf8-bb49-4df3509ab20f req-23227732-7e32-4f02-abb2-6221707a3b1f service nova] Acquired lock "refresh_cache-d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.953919] env[61970]: DEBUG nova.network.neutron [req-ec553b84-d0d0-4cf8-bb49-4df3509ab20f req-23227732-7e32-4f02-abb2-6221707a3b1f service nova] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Refreshing network info cache for port fd66c242-f49a-4324-ac18-fba690484aa7 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 598.955155] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-590a8ddd-83e3-47a6-b95e-958c531874fa {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.971021] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b4b7de5-6991-47f5-8ecf-6270d04f5c34 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.982285] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0059bc3b-6a89-4849-9e2c-2f28b045faf8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.991109] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bdf2829-b229-493e-ae25-9b762b9b7fa1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.000654] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842 could not be found. [ 599.001046] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 599.001337] env[61970]: INFO nova.compute.manager [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Took 0.05 seconds to destroy the instance on the hypervisor. [ 599.001675] env[61970]: DEBUG oslo.service.loopingcall [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.002410] env[61970]: DEBUG nova.compute.manager [-] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.004017] env[61970]: DEBUG nova.network.neutron [-] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 599.033113] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829f94ed-8124-44b0-9d5d-d75937e8dc2b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.042618] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9476496d-0e57-46d7-9266-33fa54affd2e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.056106] env[61970]: DEBUG nova.compute.provider_tree [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.058115] env[61970]: DEBUG nova.network.neutron [-] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.315374] env[61970]: DEBUG nova.network.neutron [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.482184] env[61970]: DEBUG nova.network.neutron [req-ec553b84-d0d0-4cf8-bb49-4df3509ab20f req-23227732-7e32-4f02-abb2-6221707a3b1f service nova] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.564477] env[61970]: DEBUG nova.scheduler.client.report [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.566630] env[61970]: DEBUG nova.network.neutron [-] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.630828] env[61970]: DEBUG nova.network.neutron [req-ec553b84-d0d0-4cf8-bb49-4df3509ab20f req-23227732-7e32-4f02-abb2-6221707a3b1f service nova] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.822224] env[61970]: INFO nova.compute.manager [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] [instance: 39bc8558-ce7f-41e9-b913-9b628a220ca7] Took 1.03 seconds to deallocate network for instance. [ 599.887304] env[61970]: DEBUG nova.compute.manager [req-cc79a687-e93d-41fb-9fca-42091957883f req-83e4f8a6-890d-41ec-b5eb-24b5444f140e service nova] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Received event network-vif-deleted-fd66c242-f49a-4324-ac18-fba690484aa7 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 600.069669] env[61970]: DEBUG oslo_concurrency.lockutils [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.070036] env[61970]: DEBUG nova.compute.manager [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 600.073288] env[61970]: INFO nova.compute.manager [-] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Took 1.07 seconds to deallocate network for instance. [ 600.077659] env[61970]: DEBUG oslo_concurrency.lockutils [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.348s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.077851] env[61970]: DEBUG nova.objects.instance [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Lazy-loading 'resources' on Instance uuid 8881014c-8505-4e75-8c31-e25b6a8d2cd4 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 600.081162] env[61970]: DEBUG nova.compute.claims [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.081359] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.133407] env[61970]: DEBUG oslo_concurrency.lockutils [req-ec553b84-d0d0-4cf8-bb49-4df3509ab20f req-23227732-7e32-4f02-abb2-6221707a3b1f service nova] Releasing lock "refresh_cache-d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.580318] env[61970]: DEBUG nova.compute.utils [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 600.581868] env[61970]: DEBUG nova.compute.manager [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 600.583663] env[61970]: DEBUG nova.network.neutron [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 600.644774] env[61970]: DEBUG nova.policy [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '003e650d0730442ca780813683190422', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e0cb0afcc68342a18415c589150fb91f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 600.857733] env[61970]: INFO nova.scheduler.client.report [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Deleted allocations for instance 39bc8558-ce7f-41e9-b913-9b628a220ca7 [ 600.978042] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62425cbe-9228-41c7-98fa-d5dd4b0ac08d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.986290] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd669b5d-cff9-4378-9d77-27dae0186e9e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.022331] env[61970]: DEBUG nova.network.neutron [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Successfully created port: 62aca399-f3b0-4df1-90de-4e7c15fe39f0 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 601.024552] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c24df45-bdc5-41b6-9e9c-64e68b85515c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.035498] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ed6c28-7b6f-4fd0-b57a-5fd29ad07ed1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.058352] env[61970]: DEBUG nova.compute.provider_tree [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.088207] env[61970]: DEBUG nova.compute.manager [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 601.372212] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca173004-7561-42db-bee5-6829b02deb8d tempest-VolumesAssistedSnapshotsTest-950256075 tempest-VolumesAssistedSnapshotsTest-950256075-project-member] Lock "39bc8558-ce7f-41e9-b913-9b628a220ca7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.862s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.437631] env[61970]: DEBUG nova.network.neutron [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Successfully created port: 7abce076-3827-4ef4-9885-481280b8aa3e {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 601.561417] env[61970]: DEBUG nova.scheduler.client.report [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.875133] env[61970]: DEBUG nova.compute.manager [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 602.040225] env[61970]: DEBUG nova.network.neutron [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Successfully created port: 257b76a0-34ef-4e34-8c9e-5bdf45238109 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 602.069367] env[61970]: DEBUG oslo_concurrency.lockutils [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.991s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.074936] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.416s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.099020] env[61970]: DEBUG nova.compute.manager [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 602.115111] env[61970]: INFO nova.scheduler.client.report [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Deleted allocations for instance 8881014c-8505-4e75-8c31-e25b6a8d2cd4 [ 602.149174] env[61970]: DEBUG nova.virt.hardware [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 602.149477] env[61970]: DEBUG nova.virt.hardware [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 602.149671] env[61970]: DEBUG nova.virt.hardware [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 602.149901] env[61970]: DEBUG nova.virt.hardware [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 602.150069] env[61970]: DEBUG nova.virt.hardware [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 602.150219] env[61970]: DEBUG nova.virt.hardware [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 602.150886] env[61970]: DEBUG nova.virt.hardware [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 602.151092] env[61970]: DEBUG nova.virt.hardware [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 602.151277] env[61970]: DEBUG nova.virt.hardware [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 602.151448] env[61970]: DEBUG nova.virt.hardware [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 602.151619] env[61970]: DEBUG nova.virt.hardware [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 602.152684] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6a17736-58ce-4268-8140-a61470f6099b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.165568] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a790ce5-c9b8-46cf-9c59-1d5d48713f14 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.396786] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.624532] env[61970]: DEBUG oslo_concurrency.lockutils [None req-61881fec-5e6e-4611-9807-c84af64e6b49 tempest-ServerDiagnosticsV248Test-1291761822 tempest-ServerDiagnosticsV248Test-1291761822-project-member] Lock "8881014c-8505-4e75-8c31-e25b6a8d2cd4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.877s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.976415] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f59c424-78c5-462c-bd5e-dfee0a01c78a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.986885] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a69961-caa8-4afb-b46f-91e0d9e02919 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.020383] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f96a427f-a437-4695-8ff7-262f663b3cc1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.029659] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6863cca4-d08e-49b9-8290-7471940c1b34 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.043544] env[61970]: DEBUG nova.compute.provider_tree [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.095173] env[61970]: DEBUG nova.compute.manager [req-214127ce-113b-4d44-9049-00b023bed0de req-2cd1da23-8607-4bfb-88d8-0c168c2c56a3 service nova] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Received event network-changed-62aca399-f3b0-4df1-90de-4e7c15fe39f0 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 603.095173] env[61970]: DEBUG nova.compute.manager [req-214127ce-113b-4d44-9049-00b023bed0de req-2cd1da23-8607-4bfb-88d8-0c168c2c56a3 service nova] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Refreshing instance network info cache due to event network-changed-62aca399-f3b0-4df1-90de-4e7c15fe39f0. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 603.095173] env[61970]: DEBUG oslo_concurrency.lockutils [req-214127ce-113b-4d44-9049-00b023bed0de req-2cd1da23-8607-4bfb-88d8-0c168c2c56a3 service nova] Acquiring lock "refresh_cache-44ea1800-005c-485f-969d-4bf71d58c284" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.095173] env[61970]: DEBUG oslo_concurrency.lockutils [req-214127ce-113b-4d44-9049-00b023bed0de req-2cd1da23-8607-4bfb-88d8-0c168c2c56a3 service nova] Acquired lock "refresh_cache-44ea1800-005c-485f-969d-4bf71d58c284" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.095173] env[61970]: DEBUG nova.network.neutron [req-214127ce-113b-4d44-9049-00b023bed0de req-2cd1da23-8607-4bfb-88d8-0c168c2c56a3 service nova] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Refreshing network info cache for port 62aca399-f3b0-4df1-90de-4e7c15fe39f0 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 603.268107] env[61970]: ERROR nova.compute.manager [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 62aca399-f3b0-4df1-90de-4e7c15fe39f0, please check neutron logs for more information. [ 603.268107] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 603.268107] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.268107] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 603.268107] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 603.268107] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 603.268107] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 603.268107] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 603.268107] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.268107] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 603.268107] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.268107] env[61970]: ERROR nova.compute.manager raise self.value [ 603.268107] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 603.268107] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 603.268107] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.268107] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 603.268690] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.268690] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 603.268690] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 62aca399-f3b0-4df1-90de-4e7c15fe39f0, please check neutron logs for more information. [ 603.268690] env[61970]: ERROR nova.compute.manager [ 603.268690] env[61970]: Traceback (most recent call last): [ 603.268906] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 603.268906] env[61970]: listener.cb(fileno) [ 603.268906] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.268906] env[61970]: result = function(*args, **kwargs) [ 603.268906] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 603.268906] env[61970]: return func(*args, **kwargs) [ 603.269116] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 603.269116] env[61970]: raise e [ 603.269116] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.269116] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 603.269116] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 603.269116] env[61970]: created_port_ids = self._update_ports_for_instance( [ 603.269116] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 603.269116] env[61970]: with excutils.save_and_reraise_exception(): [ 603.269116] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.269116] env[61970]: self.force_reraise() [ 603.269116] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.269116] env[61970]: raise self.value [ 603.269116] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 603.269116] env[61970]: updated_port = self._update_port( [ 603.269116] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.269116] env[61970]: _ensure_no_port_binding_failure(port) [ 603.269116] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.269116] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 603.269116] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 62aca399-f3b0-4df1-90de-4e7c15fe39f0, please check neutron logs for more information. [ 603.269116] env[61970]: Removing descriptor: 15 [ 603.269767] env[61970]: ERROR nova.compute.manager [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 62aca399-f3b0-4df1-90de-4e7c15fe39f0, please check neutron logs for more information. [ 603.269767] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Traceback (most recent call last): [ 603.269767] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 603.269767] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] yield resources [ 603.269767] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 603.269767] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] self.driver.spawn(context, instance, image_meta, [ 603.269767] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 603.269767] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] self._vmops.spawn(context, instance, image_meta, injected_files, [ 603.269767] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 603.269767] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] vm_ref = self.build_virtual_machine(instance, [ 603.269767] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 603.270110] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] vif_infos = vmwarevif.get_vif_info(self._session, [ 603.270110] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 603.270110] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] for vif in network_info: [ 603.270110] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 603.270110] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] return self._sync_wrapper(fn, *args, **kwargs) [ 603.270110] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 603.270110] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] self.wait() [ 603.270110] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 603.270110] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] self[:] = self._gt.wait() [ 603.270110] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 603.270110] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] return self._exit_event.wait() [ 603.270110] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 603.270110] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] result = hub.switch() [ 603.270546] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 603.270546] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] return self.greenlet.switch() [ 603.270546] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.270546] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] result = function(*args, **kwargs) [ 603.270546] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 603.270546] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] return func(*args, **kwargs) [ 603.270546] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 603.270546] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] raise e [ 603.270546] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.270546] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] nwinfo = self.network_api.allocate_for_instance( [ 603.270546] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 603.270546] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] created_port_ids = self._update_ports_for_instance( [ 603.270546] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 603.270943] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] with excutils.save_and_reraise_exception(): [ 603.270943] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.270943] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] self.force_reraise() [ 603.270943] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.270943] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] raise self.value [ 603.270943] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 603.270943] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] updated_port = self._update_port( [ 603.270943] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.270943] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] _ensure_no_port_binding_failure(port) [ 603.270943] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.270943] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] raise exception.PortBindingFailed(port_id=port['id']) [ 603.270943] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] nova.exception.PortBindingFailed: Binding failed for port 62aca399-f3b0-4df1-90de-4e7c15fe39f0, please check neutron logs for more information. [ 603.270943] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] [ 603.271343] env[61970]: INFO nova.compute.manager [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Terminating instance [ 603.273828] env[61970]: DEBUG oslo_concurrency.lockutils [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Acquiring lock "refresh_cache-44ea1800-005c-485f-969d-4bf71d58c284" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.547504] env[61970]: DEBUG nova.scheduler.client.report [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 603.644716] env[61970]: DEBUG nova.network.neutron [req-214127ce-113b-4d44-9049-00b023bed0de req-2cd1da23-8607-4bfb-88d8-0c168c2c56a3 service nova] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.740257] env[61970]: DEBUG nova.network.neutron [req-214127ce-113b-4d44-9049-00b023bed0de req-2cd1da23-8607-4bfb-88d8-0c168c2c56a3 service nova] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.052589] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.978s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.053258] env[61970]: ERROR nova.compute.manager [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 09574426-d427-42af-9a8a-a9d325d1154d, please check neutron logs for more information. [ 604.053258] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Traceback (most recent call last): [ 604.053258] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.053258] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] self.driver.spawn(context, instance, image_meta, [ 604.053258] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 604.053258] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.053258] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.053258] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] vm_ref = self.build_virtual_machine(instance, [ 604.053258] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.053258] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.053258] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.053942] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] for vif in network_info: [ 604.053942] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 604.053942] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] return self._sync_wrapper(fn, *args, **kwargs) [ 604.053942] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 604.053942] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] self.wait() [ 604.053942] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 604.053942] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] self[:] = self._gt.wait() [ 604.053942] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.053942] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] return self._exit_event.wait() [ 604.053942] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 604.053942] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] result = hub.switch() [ 604.053942] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 604.053942] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] return self.greenlet.switch() [ 604.054669] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.054669] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] result = function(*args, **kwargs) [ 604.054669] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 604.054669] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] return func(*args, **kwargs) [ 604.054669] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.054669] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] raise e [ 604.054669] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.054669] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] nwinfo = self.network_api.allocate_for_instance( [ 604.054669] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.054669] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] created_port_ids = self._update_ports_for_instance( [ 604.054669] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.054669] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] with excutils.save_and_reraise_exception(): [ 604.054669] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.055380] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] self.force_reraise() [ 604.055380] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.055380] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] raise self.value [ 604.055380] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.055380] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] updated_port = self._update_port( [ 604.055380] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.055380] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] _ensure_no_port_binding_failure(port) [ 604.055380] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.055380] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] raise exception.PortBindingFailed(port_id=port['id']) [ 604.055380] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] nova.exception.PortBindingFailed: Binding failed for port 09574426-d427-42af-9a8a-a9d325d1154d, please check neutron logs for more information. [ 604.055380] env[61970]: ERROR nova.compute.manager [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] [ 604.055991] env[61970]: DEBUG nova.compute.utils [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Binding failed for port 09574426-d427-42af-9a8a-a9d325d1154d, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 604.056477] env[61970]: DEBUG nova.compute.manager [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Build of instance 17f41a87-4fd0-4866-905a-2aa3d2e9691b was re-scheduled: Binding failed for port 09574426-d427-42af-9a8a-a9d325d1154d, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 604.056938] env[61970]: DEBUG nova.compute.manager [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 604.057190] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Acquiring lock "refresh_cache-17f41a87-4fd0-4866-905a-2aa3d2e9691b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.057336] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Acquired lock "refresh_cache-17f41a87-4fd0-4866-905a-2aa3d2e9691b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.057503] env[61970]: DEBUG nova.network.neutron [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 604.059836] env[61970]: DEBUG oslo_concurrency.lockutils [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.231s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.243922] env[61970]: DEBUG oslo_concurrency.lockutils [req-214127ce-113b-4d44-9049-00b023bed0de req-2cd1da23-8607-4bfb-88d8-0c168c2c56a3 service nova] Releasing lock "refresh_cache-44ea1800-005c-485f-969d-4bf71d58c284" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.243922] env[61970]: DEBUG oslo_concurrency.lockutils [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Acquired lock "refresh_cache-44ea1800-005c-485f-969d-4bf71d58c284" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.243922] env[61970]: DEBUG nova.network.neutron [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 604.594730] env[61970]: DEBUG nova.network.neutron [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.677807] env[61970]: DEBUG nova.network.neutron [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.800654] env[61970]: DEBUG nova.network.neutron [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.951690] env[61970]: DEBUG nova.network.neutron [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.009119] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff1861c-f5d7-45cf-a5de-28ddfeff6bb0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.016785] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63db36c7-3c99-44c0-b098-84a9b40e0b1a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.056414] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ed373f-72a1-45fc-881e-4f50073792f1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.065175] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-120ec4cb-d8f2-4576-a975-e472782a436f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.078567] env[61970]: DEBUG nova.compute.provider_tree [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.184989] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Releasing lock "refresh_cache-17f41a87-4fd0-4866-905a-2aa3d2e9691b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.185356] env[61970]: DEBUG nova.compute.manager [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 605.186456] env[61970]: DEBUG nova.compute.manager [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.186456] env[61970]: DEBUG nova.network.neutron [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 605.194259] env[61970]: DEBUG nova.compute.manager [req-4447c4f0-3922-4431-9883-93eb92b20bc9 req-4727225e-8c12-4740-8ee2-895fdb91c60f service nova] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Received event network-vif-deleted-62aca399-f3b0-4df1-90de-4e7c15fe39f0 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 605.221976] env[61970]: DEBUG nova.network.neutron [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.456163] env[61970]: DEBUG oslo_concurrency.lockutils [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Releasing lock "refresh_cache-44ea1800-005c-485f-969d-4bf71d58c284" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.456163] env[61970]: DEBUG nova.compute.manager [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 605.456163] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 605.456163] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-618d8679-23d0-4ba3-a931-6bf30f4637c9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.470536] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfaf0f4b-609d-4e64-882d-fbda02fe76e0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.500426] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 44ea1800-005c-485f-969d-4bf71d58c284 could not be found. [ 605.500426] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 605.500426] env[61970]: INFO nova.compute.manager [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Took 0.04 seconds to destroy the instance on the hypervisor. [ 605.500426] env[61970]: DEBUG oslo.service.loopingcall [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 605.500426] env[61970]: DEBUG nova.compute.manager [-] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.500426] env[61970]: DEBUG nova.network.neutron [-] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 605.582914] env[61970]: DEBUG nova.scheduler.client.report [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 605.615543] env[61970]: DEBUG nova.network.neutron [-] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.724338] env[61970]: DEBUG nova.network.neutron [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.094638] env[61970]: DEBUG oslo_concurrency.lockutils [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.035s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.095304] env[61970]: ERROR nova.compute.manager [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port abc054de-ef96-40b1-8b51-30bf5252429d, please check neutron logs for more information. [ 606.095304] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Traceback (most recent call last): [ 606.095304] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 606.095304] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] self.driver.spawn(context, instance, image_meta, [ 606.095304] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 606.095304] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.095304] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.095304] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] vm_ref = self.build_virtual_machine(instance, [ 606.095304] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.095304] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.095304] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.096290] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] for vif in network_info: [ 606.096290] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 606.096290] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] return self._sync_wrapper(fn, *args, **kwargs) [ 606.096290] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 606.096290] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] self.wait() [ 606.096290] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 606.096290] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] self[:] = self._gt.wait() [ 606.096290] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.096290] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] return self._exit_event.wait() [ 606.096290] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 606.096290] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] result = hub.switch() [ 606.096290] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 606.096290] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] return self.greenlet.switch() [ 606.098099] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.098099] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] result = function(*args, **kwargs) [ 606.098099] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 606.098099] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] return func(*args, **kwargs) [ 606.098099] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.098099] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] raise e [ 606.098099] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.098099] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] nwinfo = self.network_api.allocate_for_instance( [ 606.098099] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.098099] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] created_port_ids = self._update_ports_for_instance( [ 606.098099] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.098099] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] with excutils.save_and_reraise_exception(): [ 606.098099] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.098552] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] self.force_reraise() [ 606.098552] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.098552] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] raise self.value [ 606.098552] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.098552] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] updated_port = self._update_port( [ 606.098552] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.098552] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] _ensure_no_port_binding_failure(port) [ 606.098552] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.098552] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] raise exception.PortBindingFailed(port_id=port['id']) [ 606.098552] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] nova.exception.PortBindingFailed: Binding failed for port abc054de-ef96-40b1-8b51-30bf5252429d, please check neutron logs for more information. [ 606.098552] env[61970]: ERROR nova.compute.manager [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] [ 606.098887] env[61970]: DEBUG nova.compute.utils [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Binding failed for port abc054de-ef96-40b1-8b51-30bf5252429d, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 606.098972] env[61970]: DEBUG nova.compute.manager [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Build of instance 8bacc19d-2626-406a-a3fb-f1069674963c was re-scheduled: Binding failed for port abc054de-ef96-40b1-8b51-30bf5252429d, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 606.099460] env[61970]: DEBUG nova.compute.manager [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 606.099778] env[61970]: DEBUG oslo_concurrency.lockutils [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Acquiring lock "refresh_cache-8bacc19d-2626-406a-a3fb-f1069674963c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.100039] env[61970]: DEBUG oslo_concurrency.lockutils [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Acquired lock "refresh_cache-8bacc19d-2626-406a-a3fb-f1069674963c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.100219] env[61970]: DEBUG nova.network.neutron [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 606.101651] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.161s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.109932] env[61970]: INFO nova.compute.claims [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 606.227114] env[61970]: INFO nova.compute.manager [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 17f41a87-4fd0-4866-905a-2aa3d2e9691b] Took 1.04 seconds to deallocate network for instance. [ 606.643029] env[61970]: DEBUG nova.network.neutron [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.739800] env[61970]: DEBUG nova.network.neutron [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.916047] env[61970]: DEBUG nova.network.neutron [-] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.244507] env[61970]: DEBUG oslo_concurrency.lockutils [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Releasing lock "refresh_cache-8bacc19d-2626-406a-a3fb-f1069674963c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.244758] env[61970]: DEBUG nova.compute.manager [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 607.244911] env[61970]: DEBUG nova.compute.manager [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 607.244947] env[61970]: DEBUG nova.network.neutron [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 607.270151] env[61970]: DEBUG nova.network.neutron [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.275535] env[61970]: INFO nova.scheduler.client.report [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Deleted allocations for instance 17f41a87-4fd0-4866-905a-2aa3d2e9691b [ 607.420758] env[61970]: INFO nova.compute.manager [-] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Took 1.92 seconds to deallocate network for instance. [ 607.423292] env[61970]: DEBUG nova.compute.claims [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 607.423292] env[61970]: DEBUG oslo_concurrency.lockutils [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.594307] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30acbaf5-851b-49c4-9e7a-461a746f595b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.600919] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2e05c67-c7a7-43fc-81ea-fef678c66c7a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.653019] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b91d401e-24de-417b-b437-74f1a333b642 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.660663] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e6c3bf9-81ae-4308-b6ac-27f4d5c09fa4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.679327] env[61970]: DEBUG nova.compute.provider_tree [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.777933] env[61970]: DEBUG nova.network.neutron [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.787915] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69068483-94cf-4a55-bd32-c3430daf7321 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Lock "17f41a87-4fd0-4866-905a-2aa3d2e9691b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.797s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.126561] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Acquiring lock "630145f3-cd01-41fe-b9c5-7ebf44b1e6d9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.126730] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Lock "630145f3-cd01-41fe-b9c5-7ebf44b1e6d9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.182715] env[61970]: DEBUG nova.scheduler.client.report [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.279313] env[61970]: INFO nova.compute.manager [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] [instance: 8bacc19d-2626-406a-a3fb-f1069674963c] Took 1.03 seconds to deallocate network for instance. [ 608.291503] env[61970]: DEBUG nova.compute.manager [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 608.691092] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.587s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.691092] env[61970]: DEBUG nova.compute.manager [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 608.692452] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.564s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.820021] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.038033] env[61970]: DEBUG oslo_concurrency.lockutils [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Acquiring lock "915c3b6c-640b-430f-b264-40bf85c642b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.038033] env[61970]: DEBUG oslo_concurrency.lockutils [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Lock "915c3b6c-640b-430f-b264-40bf85c642b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.197628] env[61970]: DEBUG nova.compute.utils [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 609.204153] env[61970]: DEBUG nova.compute.manager [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 609.204292] env[61970]: DEBUG nova.network.neutron [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 609.283619] env[61970]: DEBUG nova.policy [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0c17548e57454c969873252035487f69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5601cf4cffea4aa594bc541a39c4c41b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 609.312615] env[61970]: INFO nova.scheduler.client.report [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Deleted allocations for instance 8bacc19d-2626-406a-a3fb-f1069674963c [ 609.615068] env[61970]: DEBUG nova.network.neutron [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Successfully created port: f763ac57-bd20-4f40-b28b-d90496239be4 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 609.618803] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022d35a0-4850-4b1f-be04-7da39c5b7bc9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.626065] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5802890-672e-4dc0-abda-5db76ecf1ce2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.662619] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea994856-d60a-475a-a6d8-9164b953f8b1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.670818] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3648ae1c-f345-4ef7-812d-5035ee914ee0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.687294] env[61970]: DEBUG nova.compute.provider_tree [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.704770] env[61970]: DEBUG nova.compute.manager [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 609.825115] env[61970]: DEBUG oslo_concurrency.lockutils [None req-995f4bbd-8cdb-48e5-ac89-fb942ade7726 tempest-ListImageFiltersTestJSON-459160492 tempest-ListImageFiltersTestJSON-459160492-project-member] Lock "8bacc19d-2626-406a-a3fb-f1069674963c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.024s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.191219] env[61970]: DEBUG nova.scheduler.client.report [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 610.332672] env[61970]: DEBUG nova.compute.manager [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 610.625427] env[61970]: DEBUG nova.compute.manager [req-d6929cea-52e9-4747-a1c9-736ac94bc2e0 req-b4104f7a-94f0-449e-a8cb-657101a52a33 service nova] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Received event network-changed-f763ac57-bd20-4f40-b28b-d90496239be4 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 610.625633] env[61970]: DEBUG nova.compute.manager [req-d6929cea-52e9-4747-a1c9-736ac94bc2e0 req-b4104f7a-94f0-449e-a8cb-657101a52a33 service nova] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Refreshing instance network info cache due to event network-changed-f763ac57-bd20-4f40-b28b-d90496239be4. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 610.625849] env[61970]: DEBUG oslo_concurrency.lockutils [req-d6929cea-52e9-4747-a1c9-736ac94bc2e0 req-b4104f7a-94f0-449e-a8cb-657101a52a33 service nova] Acquiring lock "refresh_cache-59d2c196-9230-4303-9b99-92211c53cb15" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.625990] env[61970]: DEBUG oslo_concurrency.lockutils [req-d6929cea-52e9-4747-a1c9-736ac94bc2e0 req-b4104f7a-94f0-449e-a8cb-657101a52a33 service nova] Acquired lock "refresh_cache-59d2c196-9230-4303-9b99-92211c53cb15" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.626156] env[61970]: DEBUG nova.network.neutron [req-d6929cea-52e9-4747-a1c9-736ac94bc2e0 req-b4104f7a-94f0-449e-a8cb-657101a52a33 service nova] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Refreshing network info cache for port f763ac57-bd20-4f40-b28b-d90496239be4 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 610.699602] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.007s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.700242] env[61970]: ERROR nova.compute.manager [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9ca8ea05-ac30-4e52-8de3-5ec6b39bdc8f, please check neutron logs for more information. [ 610.700242] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Traceback (most recent call last): [ 610.700242] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.700242] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] self.driver.spawn(context, instance, image_meta, [ 610.700242] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 610.700242] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.700242] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.700242] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] vm_ref = self.build_virtual_machine(instance, [ 610.700242] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.700242] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.700242] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.700568] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] for vif in network_info: [ 610.700568] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 610.700568] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] return self._sync_wrapper(fn, *args, **kwargs) [ 610.700568] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 610.700568] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] self.wait() [ 610.700568] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 610.700568] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] self[:] = self._gt.wait() [ 610.700568] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.700568] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] return self._exit_event.wait() [ 610.700568] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.700568] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] result = hub.switch() [ 610.700568] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.700568] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] return self.greenlet.switch() [ 610.700916] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.700916] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] result = function(*args, **kwargs) [ 610.700916] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.700916] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] return func(*args, **kwargs) [ 610.700916] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.700916] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] raise e [ 610.700916] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.700916] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] nwinfo = self.network_api.allocate_for_instance( [ 610.700916] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.700916] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] created_port_ids = self._update_ports_for_instance( [ 610.700916] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.700916] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] with excutils.save_and_reraise_exception(): [ 610.700916] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.701180] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] self.force_reraise() [ 610.701180] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.701180] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] raise self.value [ 610.701180] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.701180] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] updated_port = self._update_port( [ 610.701180] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.701180] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] _ensure_no_port_binding_failure(port) [ 610.701180] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.701180] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] raise exception.PortBindingFailed(port_id=port['id']) [ 610.701180] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] nova.exception.PortBindingFailed: Binding failed for port 9ca8ea05-ac30-4e52-8de3-5ec6b39bdc8f, please check neutron logs for more information. [ 610.701180] env[61970]: ERROR nova.compute.manager [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] [ 610.701392] env[61970]: DEBUG nova.compute.utils [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Binding failed for port 9ca8ea05-ac30-4e52-8de3-5ec6b39bdc8f, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 610.703041] env[61970]: DEBUG oslo_concurrency.lockutils [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.659s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.705838] env[61970]: DEBUG nova.compute.manager [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Build of instance fb0aec23-e010-4f9c-8a1d-1b31d17190b8 was re-scheduled: Binding failed for port 9ca8ea05-ac30-4e52-8de3-5ec6b39bdc8f, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 610.710455] env[61970]: DEBUG nova.compute.manager [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 610.710455] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Acquiring lock "refresh_cache-fb0aec23-e010-4f9c-8a1d-1b31d17190b8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.710455] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Acquired lock "refresh_cache-fb0aec23-e010-4f9c-8a1d-1b31d17190b8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.710455] env[61970]: DEBUG nova.network.neutron [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 610.711322] env[61970]: ERROR nova.compute.manager [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f763ac57-bd20-4f40-b28b-d90496239be4, please check neutron logs for more information. [ 610.711322] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 610.711322] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.711322] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 610.711322] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.711322] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 610.711322] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.711322] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 610.711322] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.711322] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 610.711322] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.711322] env[61970]: ERROR nova.compute.manager raise self.value [ 610.711322] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.711322] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 610.711322] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.711322] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 610.711749] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.711749] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 610.711749] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f763ac57-bd20-4f40-b28b-d90496239be4, please check neutron logs for more information. [ 610.711749] env[61970]: ERROR nova.compute.manager [ 610.711749] env[61970]: Traceback (most recent call last): [ 610.711749] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 610.711749] env[61970]: listener.cb(fileno) [ 610.711749] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.711749] env[61970]: result = function(*args, **kwargs) [ 610.711749] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.711749] env[61970]: return func(*args, **kwargs) [ 610.711749] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.711749] env[61970]: raise e [ 610.711749] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.711749] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 610.711749] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.711749] env[61970]: created_port_ids = self._update_ports_for_instance( [ 610.711749] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.711749] env[61970]: with excutils.save_and_reraise_exception(): [ 610.711749] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.711749] env[61970]: self.force_reraise() [ 610.711749] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.711749] env[61970]: raise self.value [ 610.711749] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.711749] env[61970]: updated_port = self._update_port( [ 610.711749] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.711749] env[61970]: _ensure_no_port_binding_failure(port) [ 610.711749] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.711749] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 610.712315] env[61970]: nova.exception.PortBindingFailed: Binding failed for port f763ac57-bd20-4f40-b28b-d90496239be4, please check neutron logs for more information. [ 610.712315] env[61970]: Removing descriptor: 15 [ 610.713105] env[61970]: DEBUG nova.compute.manager [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 610.749267] env[61970]: DEBUG nova.virt.hardware [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 610.749523] env[61970]: DEBUG nova.virt.hardware [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 610.749680] env[61970]: DEBUG nova.virt.hardware [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 610.749882] env[61970]: DEBUG nova.virt.hardware [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 610.750443] env[61970]: DEBUG nova.virt.hardware [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 610.751666] env[61970]: DEBUG nova.virt.hardware [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 610.751913] env[61970]: DEBUG nova.virt.hardware [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 610.752096] env[61970]: DEBUG nova.virt.hardware [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 610.752272] env[61970]: DEBUG nova.virt.hardware [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 610.752695] env[61970]: DEBUG nova.virt.hardware [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 610.754740] env[61970]: DEBUG nova.virt.hardware [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.755811] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a23c41b-3c4f-4449-adc1-0c31a0f8e651 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.765877] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3414054-90c9-4a7f-9a12-790539191f05 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.782249] env[61970]: ERROR nova.compute.manager [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f763ac57-bd20-4f40-b28b-d90496239be4, please check neutron logs for more information. [ 610.782249] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Traceback (most recent call last): [ 610.782249] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 610.782249] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] yield resources [ 610.782249] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.782249] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] self.driver.spawn(context, instance, image_meta, [ 610.782249] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 610.782249] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.782249] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.782249] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] vm_ref = self.build_virtual_machine(instance, [ 610.782249] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.782642] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.782642] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.782642] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] for vif in network_info: [ 610.782642] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 610.782642] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] return self._sync_wrapper(fn, *args, **kwargs) [ 610.782642] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 610.782642] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] self.wait() [ 610.782642] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 610.782642] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] self[:] = self._gt.wait() [ 610.782642] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.782642] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] return self._exit_event.wait() [ 610.782642] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 610.782642] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] current.throw(*self._exc) [ 610.782977] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.782977] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] result = function(*args, **kwargs) [ 610.782977] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.782977] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] return func(*args, **kwargs) [ 610.782977] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.782977] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] raise e [ 610.782977] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.782977] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] nwinfo = self.network_api.allocate_for_instance( [ 610.782977] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.782977] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] created_port_ids = self._update_ports_for_instance( [ 610.782977] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.782977] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] with excutils.save_and_reraise_exception(): [ 610.782977] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.783367] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] self.force_reraise() [ 610.783367] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.783367] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] raise self.value [ 610.783367] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.783367] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] updated_port = self._update_port( [ 610.783367] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.783367] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] _ensure_no_port_binding_failure(port) [ 610.783367] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.783367] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] raise exception.PortBindingFailed(port_id=port['id']) [ 610.783367] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] nova.exception.PortBindingFailed: Binding failed for port f763ac57-bd20-4f40-b28b-d90496239be4, please check neutron logs for more information. [ 610.783367] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] [ 610.783367] env[61970]: INFO nova.compute.manager [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Terminating instance [ 610.784440] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Acquiring lock "refresh_cache-59d2c196-9230-4303-9b99-92211c53cb15" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.853286] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.147244] env[61970]: DEBUG nova.network.neutron [req-d6929cea-52e9-4747-a1c9-736ac94bc2e0 req-b4104f7a-94f0-449e-a8cb-657101a52a33 service nova] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.233191] env[61970]: DEBUG nova.network.neutron [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.251482] env[61970]: DEBUG nova.network.neutron [req-d6929cea-52e9-4747-a1c9-736ac94bc2e0 req-b4104f7a-94f0-449e-a8cb-657101a52a33 service nova] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.339754] env[61970]: DEBUG nova.network.neutron [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.557651] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1227a9c5-0744-44ef-bfbb-9db75379c14c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.566175] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34738cd6-b639-4fb1-a4f0-0decd5689eeb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.600795] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-334c1f0b-40ef-45e3-8446-0ca69bb87372 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.608874] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a809c0-a3c0-4794-b1b0-6616327ba1be {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.622927] env[61970]: DEBUG nova.compute.provider_tree [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.754069] env[61970]: DEBUG oslo_concurrency.lockutils [req-d6929cea-52e9-4747-a1c9-736ac94bc2e0 req-b4104f7a-94f0-449e-a8cb-657101a52a33 service nova] Releasing lock "refresh_cache-59d2c196-9230-4303-9b99-92211c53cb15" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.754450] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Acquired lock "refresh_cache-59d2c196-9230-4303-9b99-92211c53cb15" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.754531] env[61970]: DEBUG nova.network.neutron [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.842185] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Releasing lock "refresh_cache-fb0aec23-e010-4f9c-8a1d-1b31d17190b8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.843044] env[61970]: DEBUG nova.compute.manager [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 611.843044] env[61970]: DEBUG nova.compute.manager [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.843044] env[61970]: DEBUG nova.network.neutron [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 611.858025] env[61970]: DEBUG nova.network.neutron [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.127156] env[61970]: DEBUG nova.scheduler.client.report [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 612.164518] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Acquiring lock "1a9257f6-5493-4ebb-8c37-88e11480a4f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.164518] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Lock "1a9257f6-5493-4ebb-8c37-88e11480a4f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.278316] env[61970]: DEBUG nova.network.neutron [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.362181] env[61970]: DEBUG nova.network.neutron [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.484781] env[61970]: DEBUG nova.network.neutron [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.632938] env[61970]: DEBUG oslo_concurrency.lockutils [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.930s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.633943] env[61970]: ERROR nova.compute.manager [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a8e611aa-9ddc-45d3-a73a-619ee9d10767, please check neutron logs for more information. [ 612.633943] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Traceback (most recent call last): [ 612.633943] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 612.633943] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] self.driver.spawn(context, instance, image_meta, [ 612.633943] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 612.633943] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 612.633943] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 612.633943] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] vm_ref = self.build_virtual_machine(instance, [ 612.633943] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 612.633943] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] vif_infos = vmwarevif.get_vif_info(self._session, [ 612.633943] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 612.634245] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] for vif in network_info: [ 612.634245] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 612.634245] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] return self._sync_wrapper(fn, *args, **kwargs) [ 612.634245] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 612.634245] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] self.wait() [ 612.634245] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 612.634245] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] self[:] = self._gt.wait() [ 612.634245] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 612.634245] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] return self._exit_event.wait() [ 612.634245] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 612.634245] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] result = hub.switch() [ 612.634245] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 612.634245] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] return self.greenlet.switch() [ 612.634507] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.634507] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] result = function(*args, **kwargs) [ 612.634507] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 612.634507] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] return func(*args, **kwargs) [ 612.634507] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 612.634507] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] raise e [ 612.634507] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 612.634507] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] nwinfo = self.network_api.allocate_for_instance( [ 612.634507] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.634507] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] created_port_ids = self._update_ports_for_instance( [ 612.634507] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.634507] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] with excutils.save_and_reraise_exception(): [ 612.634507] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.634764] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] self.force_reraise() [ 612.634764] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.634764] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] raise self.value [ 612.634764] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.634764] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] updated_port = self._update_port( [ 612.634764] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.634764] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] _ensure_no_port_binding_failure(port) [ 612.634764] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.634764] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] raise exception.PortBindingFailed(port_id=port['id']) [ 612.634764] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] nova.exception.PortBindingFailed: Binding failed for port a8e611aa-9ddc-45d3-a73a-619ee9d10767, please check neutron logs for more information. [ 612.634764] env[61970]: ERROR nova.compute.manager [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] [ 612.634987] env[61970]: DEBUG nova.compute.utils [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Binding failed for port a8e611aa-9ddc-45d3-a73a-619ee9d10767, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 612.636888] env[61970]: DEBUG nova.compute.manager [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Build of instance 0ef75e57-83af-4440-890e-95dcdb9920d2 was re-scheduled: Binding failed for port a8e611aa-9ddc-45d3-a73a-619ee9d10767, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 612.637073] env[61970]: DEBUG nova.compute.manager [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 612.637405] env[61970]: DEBUG oslo_concurrency.lockutils [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Acquiring lock "refresh_cache-0ef75e57-83af-4440-890e-95dcdb9920d2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.637663] env[61970]: DEBUG oslo_concurrency.lockutils [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Acquired lock "refresh_cache-0ef75e57-83af-4440-890e-95dcdb9920d2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.637887] env[61970]: DEBUG nova.network.neutron [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.639319] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.541s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.646392] env[61970]: DEBUG nova.compute.manager [req-0c686b01-3a9c-49fc-b892-c18136eff386 req-ad679833-30dc-4a92-88bc-57b901a35cc1 service nova] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Received event network-vif-deleted-f763ac57-bd20-4f40-b28b-d90496239be4 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 612.864621] env[61970]: INFO nova.compute.manager [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] [instance: fb0aec23-e010-4f9c-8a1d-1b31d17190b8] Took 1.02 seconds to deallocate network for instance. [ 612.988263] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Releasing lock "refresh_cache-59d2c196-9230-4303-9b99-92211c53cb15" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.988626] env[61970]: DEBUG nova.compute.manager [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 612.988817] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 612.989132] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a2cfd40c-5bff-4219-a7d4-e87eb3f70ad7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.998403] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b12772c-9d0e-4c4f-b185-4f1fc51c6f7b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.019840] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 59d2c196-9230-4303-9b99-92211c53cb15 could not be found. [ 613.020055] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 613.020237] env[61970]: INFO nova.compute.manager [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Took 0.03 seconds to destroy the instance on the hypervisor. [ 613.020475] env[61970]: DEBUG oslo.service.loopingcall [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 613.020675] env[61970]: DEBUG nova.compute.manager [-] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.020768] env[61970]: DEBUG nova.network.neutron [-] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.040488] env[61970]: DEBUG nova.network.neutron [-] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.164995] env[61970]: DEBUG nova.network.neutron [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.247230] env[61970]: DEBUG nova.network.neutron [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.480151] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5ef2ee7-4935-4db8-a2f5-a0c385750ccb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.487719] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9796c562-c5f3-4940-bff4-f015c9f8e6dc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.517893] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f688e97-a6df-4448-b1dc-7b730cc57fc4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.525425] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-463187df-22d2-4d30-9ec8-480ea95c4267 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.538628] env[61970]: DEBUG nova.compute.provider_tree [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.543162] env[61970]: DEBUG nova.network.neutron [-] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.750043] env[61970]: DEBUG oslo_concurrency.lockutils [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Releasing lock "refresh_cache-0ef75e57-83af-4440-890e-95dcdb9920d2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.750292] env[61970]: DEBUG nova.compute.manager [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 613.750458] env[61970]: DEBUG nova.compute.manager [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.750625] env[61970]: DEBUG nova.network.neutron [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.768084] env[61970]: DEBUG nova.network.neutron [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.893761] env[61970]: INFO nova.scheduler.client.report [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Deleted allocations for instance fb0aec23-e010-4f9c-8a1d-1b31d17190b8 [ 614.041946] env[61970]: DEBUG nova.scheduler.client.report [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.046507] env[61970]: INFO nova.compute.manager [-] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Took 1.03 seconds to deallocate network for instance. [ 614.049413] env[61970]: DEBUG nova.compute.claims [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 614.049529] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.270677] env[61970]: DEBUG nova.network.neutron [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.404964] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69b1968b-8036-46cc-b75b-9c5583f20547 tempest-FloatingIPsAssociationTestJSON-1998467955 tempest-FloatingIPsAssociationTestJSON-1998467955-project-member] Lock "fb0aec23-e010-4f9c-8a1d-1b31d17190b8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.892s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.552025] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.910s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.552025] env[61970]: ERROR nova.compute.manager [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a4134d35-8c9e-43b0-8491-44c9fc75a801, please check neutron logs for more information. [ 614.552025] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Traceback (most recent call last): [ 614.552025] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 614.552025] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] self.driver.spawn(context, instance, image_meta, [ 614.552025] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 614.552025] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.552025] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.552025] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] vm_ref = self.build_virtual_machine(instance, [ 614.552521] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.552521] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.552521] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.552521] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] for vif in network_info: [ 614.552521] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 614.552521] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] return self._sync_wrapper(fn, *args, **kwargs) [ 614.552521] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 614.552521] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] self.wait() [ 614.552521] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 614.552521] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] self[:] = self._gt.wait() [ 614.552521] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.552521] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] return self._exit_event.wait() [ 614.552521] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 614.552835] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] result = hub.switch() [ 614.552835] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 614.552835] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] return self.greenlet.switch() [ 614.552835] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.552835] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] result = function(*args, **kwargs) [ 614.552835] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 614.552835] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] return func(*args, **kwargs) [ 614.552835] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.552835] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] raise e [ 614.552835] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.552835] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] nwinfo = self.network_api.allocate_for_instance( [ 614.552835] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.552835] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] created_port_ids = self._update_ports_for_instance( [ 614.553251] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.553251] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] with excutils.save_and_reraise_exception(): [ 614.553251] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.553251] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] self.force_reraise() [ 614.553251] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.553251] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] raise self.value [ 614.553251] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.553251] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] updated_port = self._update_port( [ 614.553251] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.553251] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] _ensure_no_port_binding_failure(port) [ 614.553251] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.553251] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] raise exception.PortBindingFailed(port_id=port['id']) [ 614.553575] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] nova.exception.PortBindingFailed: Binding failed for port a4134d35-8c9e-43b0-8491-44c9fc75a801, please check neutron logs for more information. [ 614.553575] env[61970]: ERROR nova.compute.manager [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] [ 614.553575] env[61970]: DEBUG nova.compute.utils [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Binding failed for port a4134d35-8c9e-43b0-8491-44c9fc75a801, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 614.553575] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.660s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.554096] env[61970]: INFO nova.compute.claims [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.558485] env[61970]: DEBUG nova.compute.manager [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Build of instance de9ed575-9e43-4c67-89ed-0549b43fb81a was re-scheduled: Binding failed for port a4134d35-8c9e-43b0-8491-44c9fc75a801, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 614.558906] env[61970]: DEBUG nova.compute.manager [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 614.559138] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Acquiring lock "refresh_cache-de9ed575-9e43-4c67-89ed-0549b43fb81a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.559312] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Acquired lock "refresh_cache-de9ed575-9e43-4c67-89ed-0549b43fb81a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.559436] env[61970]: DEBUG nova.network.neutron [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 614.773787] env[61970]: INFO nova.compute.manager [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] [instance: 0ef75e57-83af-4440-890e-95dcdb9920d2] Took 1.02 seconds to deallocate network for instance. [ 614.909012] env[61970]: DEBUG nova.compute.manager [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 615.081536] env[61970]: DEBUG nova.network.neutron [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.197201] env[61970]: DEBUG nova.network.neutron [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.439634] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.699588] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Releasing lock "refresh_cache-de9ed575-9e43-4c67-89ed-0549b43fb81a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.699840] env[61970]: DEBUG nova.compute.manager [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 615.700037] env[61970]: DEBUG nova.compute.manager [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.700210] env[61970]: DEBUG nova.network.neutron [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 615.740378] env[61970]: DEBUG nova.network.neutron [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.803354] env[61970]: INFO nova.scheduler.client.report [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Deleted allocations for instance 0ef75e57-83af-4440-890e-95dcdb9920d2 [ 616.019759] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0d3661-b866-4724-9a07-91af1f7bcf7b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.027532] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e81b173-06b9-4ef4-a3c3-68af846cc851 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.071710] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb37d062-1bf5-44f8-a4f9-a47eb6f28823 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.080185] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c7d0178-4282-427e-900c-f04c53a7104b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.097610] env[61970]: DEBUG nova.compute.provider_tree [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.106986] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Acquiring lock "762efd5e-c95e-4718-98de-2e0b05226c06" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.107206] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Lock "762efd5e-c95e-4718-98de-2e0b05226c06" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.246383] env[61970]: DEBUG nova.network.neutron [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.316282] env[61970]: DEBUG oslo_concurrency.lockutils [None req-57c76922-887b-482d-a539-f7b2722408a9 tempest-TenantUsagesTestJSON-939833436 tempest-TenantUsagesTestJSON-939833436-project-member] Lock "0ef75e57-83af-4440-890e-95dcdb9920d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.721s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.601593] env[61970]: DEBUG nova.scheduler.client.report [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.750772] env[61970]: INFO nova.compute.manager [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] [instance: de9ed575-9e43-4c67-89ed-0549b43fb81a] Took 1.05 seconds to deallocate network for instance. [ 616.816242] env[61970]: DEBUG nova.compute.manager [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 617.110007] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.558s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.110590] env[61970]: DEBUG nova.compute.manager [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 617.116197] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.035s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.339889] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.618108] env[61970]: DEBUG nova.compute.utils [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 617.621127] env[61970]: DEBUG nova.compute.manager [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 617.621127] env[61970]: DEBUG nova.network.neutron [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 617.664871] env[61970]: DEBUG nova.policy [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '13aacb75013d408980ba5189c6071447', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c5b87b23cfe747379a4f42e6879f0f25', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 617.789058] env[61970]: INFO nova.scheduler.client.report [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Deleted allocations for instance de9ed575-9e43-4c67-89ed-0549b43fb81a [ 618.028509] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d0fd6e-be69-4aad-b5ac-37a3b567caf2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.037664] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee191cbf-25c0-47f1-9407-6ab89e188e98 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.073360] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-057713cd-6674-459e-a0b0-a3f3bba40b4a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.076600] env[61970]: DEBUG nova.network.neutron [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Successfully created port: d01aba36-d853-4a23-b25f-dedaf27edf79 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 618.081616] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57283dc2-b506-4ec5-bbee-152d5f7e8680 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.099226] env[61970]: DEBUG nova.compute.provider_tree [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.121603] env[61970]: DEBUG nova.compute.manager [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 618.297834] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f5e601fb-e0c8-4e74-9ff5-be6c4e280405 tempest-ImagesNegativeTestJSON-297618587 tempest-ImagesNegativeTestJSON-297618587-project-member] Lock "de9ed575-9e43-4c67-89ed-0549b43fb81a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.322s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.602509] env[61970]: DEBUG nova.scheduler.client.report [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.802659] env[61970]: DEBUG nova.compute.manager [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 619.110793] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.992s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.110793] env[61970]: ERROR nova.compute.manager [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fd66c242-f49a-4324-ac18-fba690484aa7, please check neutron logs for more information. [ 619.110793] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Traceback (most recent call last): [ 619.110793] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 619.110793] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] self.driver.spawn(context, instance, image_meta, [ 619.110793] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 619.110793] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.110793] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.110793] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] vm_ref = self.build_virtual_machine(instance, [ 619.111221] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.111221] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.111221] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.111221] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] for vif in network_info: [ 619.111221] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 619.111221] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] return self._sync_wrapper(fn, *args, **kwargs) [ 619.111221] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 619.111221] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] self.wait() [ 619.111221] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 619.111221] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] self[:] = self._gt.wait() [ 619.111221] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.111221] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] return self._exit_event.wait() [ 619.111221] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 619.111518] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] result = hub.switch() [ 619.111518] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 619.111518] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] return self.greenlet.switch() [ 619.111518] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.111518] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] result = function(*args, **kwargs) [ 619.111518] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 619.111518] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] return func(*args, **kwargs) [ 619.111518] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.111518] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] raise e [ 619.111518] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.111518] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] nwinfo = self.network_api.allocate_for_instance( [ 619.111518] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.111518] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] created_port_ids = self._update_ports_for_instance( [ 619.111816] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.111816] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] with excutils.save_and_reraise_exception(): [ 619.111816] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.111816] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] self.force_reraise() [ 619.111816] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.111816] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] raise self.value [ 619.111816] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.111816] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] updated_port = self._update_port( [ 619.111816] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.111816] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] _ensure_no_port_binding_failure(port) [ 619.111816] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.111816] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] raise exception.PortBindingFailed(port_id=port['id']) [ 619.112111] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] nova.exception.PortBindingFailed: Binding failed for port fd66c242-f49a-4324-ac18-fba690484aa7, please check neutron logs for more information. [ 619.112111] env[61970]: ERROR nova.compute.manager [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] [ 619.112111] env[61970]: DEBUG nova.compute.utils [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Binding failed for port fd66c242-f49a-4324-ac18-fba690484aa7, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 619.113035] env[61970]: DEBUG nova.compute.manager [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Build of instance d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842 was re-scheduled: Binding failed for port fd66c242-f49a-4324-ac18-fba690484aa7, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 619.113960] env[61970]: DEBUG nova.compute.manager [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 619.114322] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Acquiring lock "refresh_cache-d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.117101] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Acquired lock "refresh_cache-d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.117101] env[61970]: DEBUG nova.network.neutron [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.117101] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.719s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.117992] env[61970]: INFO nova.compute.claims [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.132655] env[61970]: DEBUG nova.compute.manager [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 619.174379] env[61970]: DEBUG nova.virt.hardware [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 619.174837] env[61970]: DEBUG nova.virt.hardware [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 619.175116] env[61970]: DEBUG nova.virt.hardware [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 619.175446] env[61970]: DEBUG nova.virt.hardware [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 619.175693] env[61970]: DEBUG nova.virt.hardware [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 619.175929] env[61970]: DEBUG nova.virt.hardware [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 619.176240] env[61970]: DEBUG nova.virt.hardware [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 619.176543] env[61970]: DEBUG nova.virt.hardware [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 619.176827] env[61970]: DEBUG nova.virt.hardware [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 619.177098] env[61970]: DEBUG nova.virt.hardware [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 619.177371] env[61970]: DEBUG nova.virt.hardware [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 619.178382] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-274dee20-a7bb-440c-983e-086d8323597c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.192019] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55150755-0ae9-4ff1-925c-ebb34978bf4b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.328937] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.662985] env[61970]: DEBUG nova.network.neutron [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.943341] env[61970]: DEBUG nova.network.neutron [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.175833] env[61970]: DEBUG nova.compute.manager [req-545c9dd2-dbb0-456f-bb0d-9ebb89de4614 req-94ee6c66-5f16-4ca9-83fd-06122624e9ba service nova] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Received event network-changed-d01aba36-d853-4a23-b25f-dedaf27edf79 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 620.177534] env[61970]: DEBUG nova.compute.manager [req-545c9dd2-dbb0-456f-bb0d-9ebb89de4614 req-94ee6c66-5f16-4ca9-83fd-06122624e9ba service nova] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Refreshing instance network info cache due to event network-changed-d01aba36-d853-4a23-b25f-dedaf27edf79. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 620.177824] env[61970]: DEBUG oslo_concurrency.lockutils [req-545c9dd2-dbb0-456f-bb0d-9ebb89de4614 req-94ee6c66-5f16-4ca9-83fd-06122624e9ba service nova] Acquiring lock "refresh_cache-7c48e5b6-eb5f-4c35-9966-1cfc69182595" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.177931] env[61970]: DEBUG oslo_concurrency.lockutils [req-545c9dd2-dbb0-456f-bb0d-9ebb89de4614 req-94ee6c66-5f16-4ca9-83fd-06122624e9ba service nova] Acquired lock "refresh_cache-7c48e5b6-eb5f-4c35-9966-1cfc69182595" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.178106] env[61970]: DEBUG nova.network.neutron [req-545c9dd2-dbb0-456f-bb0d-9ebb89de4614 req-94ee6c66-5f16-4ca9-83fd-06122624e9ba service nova] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Refreshing network info cache for port d01aba36-d853-4a23-b25f-dedaf27edf79 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 620.448959] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Releasing lock "refresh_cache-d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.449241] env[61970]: DEBUG nova.compute.manager [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 620.449447] env[61970]: DEBUG nova.compute.manager [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.449618] env[61970]: DEBUG nova.network.neutron [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 620.492198] env[61970]: DEBUG nova.network.neutron [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.540881] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f2fa8d-3301-43fe-b21f-37666d7ed6ef {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.547684] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1588e7b3-3cc5-4b86-b965-d0625909cc60 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.586051] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6580c57b-423e-4ad3-b17c-bcc43c8909c4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.594724] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2eaa92-764b-4531-8587-088f816869cc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.617196] env[61970]: DEBUG nova.compute.provider_tree [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.731018] env[61970]: DEBUG nova.network.neutron [req-545c9dd2-dbb0-456f-bb0d-9ebb89de4614 req-94ee6c66-5f16-4ca9-83fd-06122624e9ba service nova] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.836418] env[61970]: ERROR nova.compute.manager [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d01aba36-d853-4a23-b25f-dedaf27edf79, please check neutron logs for more information. [ 620.836418] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 620.836418] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.836418] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 620.836418] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.836418] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 620.836418] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.836418] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 620.836418] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.836418] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 620.836418] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.836418] env[61970]: ERROR nova.compute.manager raise self.value [ 620.836418] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.836418] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 620.836418] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.836418] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 620.836828] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.836828] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 620.836828] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d01aba36-d853-4a23-b25f-dedaf27edf79, please check neutron logs for more information. [ 620.836828] env[61970]: ERROR nova.compute.manager [ 620.836828] env[61970]: Traceback (most recent call last): [ 620.836828] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 620.836828] env[61970]: listener.cb(fileno) [ 620.836828] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.836828] env[61970]: result = function(*args, **kwargs) [ 620.836828] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 620.836828] env[61970]: return func(*args, **kwargs) [ 620.836828] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 620.836828] env[61970]: raise e [ 620.836828] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.836828] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 620.836828] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.836828] env[61970]: created_port_ids = self._update_ports_for_instance( [ 620.836828] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.836828] env[61970]: with excutils.save_and_reraise_exception(): [ 620.836828] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.836828] env[61970]: self.force_reraise() [ 620.836828] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.836828] env[61970]: raise self.value [ 620.836828] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.836828] env[61970]: updated_port = self._update_port( [ 620.836828] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.836828] env[61970]: _ensure_no_port_binding_failure(port) [ 620.836828] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.836828] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 620.837489] env[61970]: nova.exception.PortBindingFailed: Binding failed for port d01aba36-d853-4a23-b25f-dedaf27edf79, please check neutron logs for more information. [ 620.837489] env[61970]: Removing descriptor: 17 [ 620.837489] env[61970]: ERROR nova.compute.manager [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d01aba36-d853-4a23-b25f-dedaf27edf79, please check neutron logs for more information. [ 620.837489] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Traceback (most recent call last): [ 620.837489] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 620.837489] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] yield resources [ 620.837489] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 620.837489] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] self.driver.spawn(context, instance, image_meta, [ 620.837489] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 620.837489] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] self._vmops.spawn(context, instance, image_meta, injected_files, [ 620.837489] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 620.837489] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] vm_ref = self.build_virtual_machine(instance, [ 620.837751] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 620.837751] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] vif_infos = vmwarevif.get_vif_info(self._session, [ 620.837751] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 620.837751] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] for vif in network_info: [ 620.837751] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 620.837751] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] return self._sync_wrapper(fn, *args, **kwargs) [ 620.837751] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 620.837751] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] self.wait() [ 620.837751] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 620.837751] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] self[:] = self._gt.wait() [ 620.837751] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 620.837751] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] return self._exit_event.wait() [ 620.837751] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 620.838145] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] result = hub.switch() [ 620.838145] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 620.838145] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] return self.greenlet.switch() [ 620.838145] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.838145] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] result = function(*args, **kwargs) [ 620.838145] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 620.838145] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] return func(*args, **kwargs) [ 620.838145] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 620.838145] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] raise e [ 620.838145] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.838145] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] nwinfo = self.network_api.allocate_for_instance( [ 620.838145] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.838145] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] created_port_ids = self._update_ports_for_instance( [ 620.838454] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.838454] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] with excutils.save_and_reraise_exception(): [ 620.838454] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.838454] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] self.force_reraise() [ 620.838454] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.838454] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] raise self.value [ 620.838454] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.838454] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] updated_port = self._update_port( [ 620.838454] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.838454] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] _ensure_no_port_binding_failure(port) [ 620.838454] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.838454] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] raise exception.PortBindingFailed(port_id=port['id']) [ 620.838729] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] nova.exception.PortBindingFailed: Binding failed for port d01aba36-d853-4a23-b25f-dedaf27edf79, please check neutron logs for more information. [ 620.838729] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] [ 620.838729] env[61970]: INFO nova.compute.manager [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Terminating instance [ 620.840389] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "refresh_cache-7c48e5b6-eb5f-4c35-9966-1cfc69182595" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.967472] env[61970]: DEBUG nova.network.neutron [req-545c9dd2-dbb0-456f-bb0d-9ebb89de4614 req-94ee6c66-5f16-4ca9-83fd-06122624e9ba service nova] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.995893] env[61970]: DEBUG nova.network.neutron [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.120361] env[61970]: DEBUG nova.scheduler.client.report [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 621.470600] env[61970]: DEBUG oslo_concurrency.lockutils [req-545c9dd2-dbb0-456f-bb0d-9ebb89de4614 req-94ee6c66-5f16-4ca9-83fd-06122624e9ba service nova] Releasing lock "refresh_cache-7c48e5b6-eb5f-4c35-9966-1cfc69182595" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.471027] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquired lock "refresh_cache-7c48e5b6-eb5f-4c35-9966-1cfc69182595" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.471218] env[61970]: DEBUG nova.network.neutron [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.498587] env[61970]: INFO nova.compute.manager [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] [instance: d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842] Took 1.05 seconds to deallocate network for instance. [ 621.627323] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.627674] env[61970]: DEBUG nova.compute.manager [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 621.631300] env[61970]: DEBUG oslo_concurrency.lockutils [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.207s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.996839] env[61970]: DEBUG nova.network.neutron [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.093141] env[61970]: DEBUG nova.network.neutron [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.136557] env[61970]: DEBUG nova.compute.utils [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 622.143670] env[61970]: DEBUG nova.compute.manager [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 622.144535] env[61970]: DEBUG nova.network.neutron [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 622.213945] env[61970]: DEBUG nova.policy [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2cbeccb03a64b8ba57d30a4d82b1931', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04e68079ffc949f09bd1b9b72ba632f1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 622.466708] env[61970]: DEBUG nova.compute.manager [req-8d31d0fd-ff6b-4548-bce7-c45a30eeed95 req-e2bf2ed9-9fb3-4837-a374-f1fa12c05ba0 service nova] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Received event network-vif-deleted-d01aba36-d853-4a23-b25f-dedaf27edf79 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 622.537223] env[61970]: INFO nova.scheduler.client.report [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Deleted allocations for instance d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842 [ 622.564738] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6873c96d-078c-4274-971f-f72eb644d75c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.575439] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22387c75-47d8-46d8-a628-f2c1e874866e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.608175] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Releasing lock "refresh_cache-7c48e5b6-eb5f-4c35-9966-1cfc69182595" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.608637] env[61970]: DEBUG nova.compute.manager [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 622.609321] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 622.610172] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b2fbe527-d0f3-433c-82eb-3b8909cd0fb6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.612563] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2211db1-36d7-4eb4-866f-a432793820c6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.622151] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1af22e-5250-40f8-8e1c-efbd0909d2b7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.628929] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31dbd4c3-3b50-456e-bbe8-6750f52db5b7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.641511] env[61970]: DEBUG nova.network.neutron [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Successfully created port: f3160862-d364-49dd-8639-282b45d418b5 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 622.651656] env[61970]: DEBUG nova.compute.manager [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 622.654503] env[61970]: DEBUG nova.compute.provider_tree [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.659166] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7c48e5b6-eb5f-4c35-9966-1cfc69182595 could not be found. [ 622.659348] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 622.659527] env[61970]: INFO nova.compute.manager [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Took 0.05 seconds to destroy the instance on the hypervisor. [ 622.659770] env[61970]: DEBUG oslo.service.loopingcall [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 622.659976] env[61970]: DEBUG nova.compute.manager [-] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.660553] env[61970]: DEBUG nova.network.neutron [-] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.675469] env[61970]: DEBUG nova.network.neutron [-] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.058846] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4135a61a-458b-41d5-835c-e3e7740830e5 tempest-ServerAddressesNegativeTestJSON-1146556747 tempest-ServerAddressesNegativeTestJSON-1146556747-project-member] Lock "d6d6e8c7-75d7-4fc5-bbdf-0af629d7d842" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.933s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.164933] env[61970]: DEBUG nova.scheduler.client.report [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.177844] env[61970]: DEBUG nova.network.neutron [-] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.503553] env[61970]: DEBUG oslo_concurrency.lockutils [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Acquiring lock "a540095e-563e-4059-b861-5d0e1b4995f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.504655] env[61970]: DEBUG oslo_concurrency.lockutils [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Lock "a540095e-563e-4059-b861-5d0e1b4995f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.562754] env[61970]: DEBUG nova.compute.manager [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 623.662053] env[61970]: DEBUG nova.compute.manager [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 623.676169] env[61970]: DEBUG oslo_concurrency.lockutils [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.043s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.676169] env[61970]: ERROR nova.compute.manager [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 62aca399-f3b0-4df1-90de-4e7c15fe39f0, please check neutron logs for more information. [ 623.676169] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Traceback (most recent call last): [ 623.676169] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 623.676169] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] self.driver.spawn(context, instance, image_meta, [ 623.676169] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 623.676169] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.676169] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.676169] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] vm_ref = self.build_virtual_machine(instance, [ 623.676434] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.676434] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.676434] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.676434] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] for vif in network_info: [ 623.676434] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 623.676434] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] return self._sync_wrapper(fn, *args, **kwargs) [ 623.676434] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 623.676434] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] self.wait() [ 623.676434] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 623.676434] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] self[:] = self._gt.wait() [ 623.676434] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.676434] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] return self._exit_event.wait() [ 623.676434] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 623.676733] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] result = hub.switch() [ 623.676733] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 623.676733] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] return self.greenlet.switch() [ 623.676733] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.676733] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] result = function(*args, **kwargs) [ 623.676733] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 623.676733] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] return func(*args, **kwargs) [ 623.676733] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.676733] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] raise e [ 623.676733] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.676733] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] nwinfo = self.network_api.allocate_for_instance( [ 623.676733] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.676733] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] created_port_ids = self._update_ports_for_instance( [ 623.677009] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.677009] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] with excutils.save_and_reraise_exception(): [ 623.677009] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.677009] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] self.force_reraise() [ 623.677009] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.677009] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] raise self.value [ 623.677009] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.677009] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] updated_port = self._update_port( [ 623.677009] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.677009] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] _ensure_no_port_binding_failure(port) [ 623.677009] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.677009] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] raise exception.PortBindingFailed(port_id=port['id']) [ 623.677266] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] nova.exception.PortBindingFailed: Binding failed for port 62aca399-f3b0-4df1-90de-4e7c15fe39f0, please check neutron logs for more information. [ 623.677266] env[61970]: ERROR nova.compute.manager [instance: 44ea1800-005c-485f-969d-4bf71d58c284] [ 623.677266] env[61970]: DEBUG nova.compute.utils [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Binding failed for port 62aca399-f3b0-4df1-90de-4e7c15fe39f0, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 623.677266] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.856s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.681489] env[61970]: INFO nova.compute.claims [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 623.686496] env[61970]: DEBUG nova.compute.manager [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Build of instance 44ea1800-005c-485f-969d-4bf71d58c284 was re-scheduled: Binding failed for port 62aca399-f3b0-4df1-90de-4e7c15fe39f0, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 623.690370] env[61970]: DEBUG nova.compute.manager [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 623.690606] env[61970]: DEBUG oslo_concurrency.lockutils [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Acquiring lock "refresh_cache-44ea1800-005c-485f-969d-4bf71d58c284" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.690749] env[61970]: DEBUG oslo_concurrency.lockutils [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Acquired lock "refresh_cache-44ea1800-005c-485f-969d-4bf71d58c284" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.690958] env[61970]: DEBUG nova.network.neutron [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 623.692824] env[61970]: INFO nova.compute.manager [-] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Took 1.03 seconds to deallocate network for instance. [ 623.696806] env[61970]: DEBUG nova.compute.claims [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 623.696949] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.714023] env[61970]: DEBUG nova.virt.hardware [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 623.714023] env[61970]: DEBUG nova.virt.hardware [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 623.714023] env[61970]: DEBUG nova.virt.hardware [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 623.714208] env[61970]: DEBUG nova.virt.hardware [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 623.714208] env[61970]: DEBUG nova.virt.hardware [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 623.714208] env[61970]: DEBUG nova.virt.hardware [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 623.714208] env[61970]: DEBUG nova.virt.hardware [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 623.714208] env[61970]: DEBUG nova.virt.hardware [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 623.714339] env[61970]: DEBUG nova.virt.hardware [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 623.714339] env[61970]: DEBUG nova.virt.hardware [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 623.714339] env[61970]: DEBUG nova.virt.hardware [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 623.714841] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db853803-585b-447e-a059-ba566f26666e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.727182] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214aea1a-6f07-43a1-82e8-94a59b2c0ce7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.002099] env[61970]: ERROR nova.compute.manager [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f3160862-d364-49dd-8639-282b45d418b5, please check neutron logs for more information. [ 624.002099] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 624.002099] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.002099] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 624.002099] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.002099] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 624.002099] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.002099] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 624.002099] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.002099] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 624.002099] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.002099] env[61970]: ERROR nova.compute.manager raise self.value [ 624.002099] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.002099] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 624.002099] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.002099] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 624.002639] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.002639] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 624.002639] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f3160862-d364-49dd-8639-282b45d418b5, please check neutron logs for more information. [ 624.002639] env[61970]: ERROR nova.compute.manager [ 624.003391] env[61970]: Traceback (most recent call last): [ 624.003428] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 624.003428] env[61970]: listener.cb(fileno) [ 624.003428] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.003428] env[61970]: result = function(*args, **kwargs) [ 624.003428] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 624.003428] env[61970]: return func(*args, **kwargs) [ 624.003428] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.003428] env[61970]: raise e [ 624.003428] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.003428] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 624.003428] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.003428] env[61970]: created_port_ids = self._update_ports_for_instance( [ 624.003428] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.003428] env[61970]: with excutils.save_and_reraise_exception(): [ 624.003428] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.003428] env[61970]: self.force_reraise() [ 624.003428] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.003428] env[61970]: raise self.value [ 624.003428] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.003428] env[61970]: updated_port = self._update_port( [ 624.003428] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.003428] env[61970]: _ensure_no_port_binding_failure(port) [ 624.003428] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.003428] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 624.003428] env[61970]: nova.exception.PortBindingFailed: Binding failed for port f3160862-d364-49dd-8639-282b45d418b5, please check neutron logs for more information. [ 624.004055] env[61970]: Removing descriptor: 15 [ 624.004484] env[61970]: ERROR nova.compute.manager [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f3160862-d364-49dd-8639-282b45d418b5, please check neutron logs for more information. [ 624.004484] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Traceback (most recent call last): [ 624.004484] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 624.004484] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] yield resources [ 624.004484] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.004484] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] self.driver.spawn(context, instance, image_meta, [ 624.004484] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 624.004484] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.004484] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.004484] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] vm_ref = self.build_virtual_machine(instance, [ 624.004484] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.004809] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.004809] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.004809] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] for vif in network_info: [ 624.004809] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 624.004809] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] return self._sync_wrapper(fn, *args, **kwargs) [ 624.004809] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 624.004809] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] self.wait() [ 624.004809] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 624.004809] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] self[:] = self._gt.wait() [ 624.004809] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.004809] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] return self._exit_event.wait() [ 624.004809] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.004809] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] result = hub.switch() [ 624.005153] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.005153] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] return self.greenlet.switch() [ 624.005153] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.005153] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] result = function(*args, **kwargs) [ 624.005153] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 624.005153] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] return func(*args, **kwargs) [ 624.005153] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.005153] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] raise e [ 624.005153] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.005153] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] nwinfo = self.network_api.allocate_for_instance( [ 624.005153] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.005153] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] created_port_ids = self._update_ports_for_instance( [ 624.005153] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.005538] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] with excutils.save_and_reraise_exception(): [ 624.005538] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.005538] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] self.force_reraise() [ 624.005538] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.005538] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] raise self.value [ 624.005538] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.005538] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] updated_port = self._update_port( [ 624.005538] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.005538] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] _ensure_no_port_binding_failure(port) [ 624.005538] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.005538] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] raise exception.PortBindingFailed(port_id=port['id']) [ 624.005538] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] nova.exception.PortBindingFailed: Binding failed for port f3160862-d364-49dd-8639-282b45d418b5, please check neutron logs for more information. [ 624.005538] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] [ 624.005875] env[61970]: INFO nova.compute.manager [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Terminating instance [ 624.009060] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Acquiring lock "refresh_cache-a6a3b8fe-6911-4d8f-bd3e-8044c47a541b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.009285] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Acquired lock "refresh_cache-a6a3b8fe-6911-4d8f-bd3e-8044c47a541b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.009500] env[61970]: DEBUG nova.network.neutron [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 624.102578] env[61970]: DEBUG oslo_concurrency.lockutils [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.213452] env[61970]: DEBUG nova.network.neutron [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.337287] env[61970]: DEBUG nova.network.neutron [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.541041] env[61970]: DEBUG nova.network.neutron [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.664668] env[61970]: DEBUG nova.network.neutron [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.723669] env[61970]: DEBUG nova.compute.manager [req-8efb9581-5d73-4939-aa5d-a1edce17265f req-53b0870f-0920-4b9a-8673-e87f941dc338 service nova] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Received event network-changed-f3160862-d364-49dd-8639-282b45d418b5 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 624.723849] env[61970]: DEBUG nova.compute.manager [req-8efb9581-5d73-4939-aa5d-a1edce17265f req-53b0870f-0920-4b9a-8673-e87f941dc338 service nova] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Refreshing instance network info cache due to event network-changed-f3160862-d364-49dd-8639-282b45d418b5. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 624.724036] env[61970]: DEBUG oslo_concurrency.lockutils [req-8efb9581-5d73-4939-aa5d-a1edce17265f req-53b0870f-0920-4b9a-8673-e87f941dc338 service nova] Acquiring lock "refresh_cache-a6a3b8fe-6911-4d8f-bd3e-8044c47a541b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.841047] env[61970]: DEBUG oslo_concurrency.lockutils [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Releasing lock "refresh_cache-44ea1800-005c-485f-969d-4bf71d58c284" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.841534] env[61970]: DEBUG nova.compute.manager [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 624.841534] env[61970]: DEBUG nova.compute.manager [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 624.841637] env[61970]: DEBUG nova.network.neutron [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 624.858216] env[61970]: DEBUG nova.network.neutron [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.071896] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4486dff-4bb6-4321-9ef4-b75d9acec8c1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.081063] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d3ddce-b45f-4a06-ad64-dcf55afe55b0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.118254] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf56d11-e061-490c-96cb-476a99ad9c9b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.126604] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a36e2c7d-93bc-4d6b-8122-396ad54c3c3f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.145704] env[61970]: DEBUG nova.compute.provider_tree [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.167610] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Releasing lock "refresh_cache-a6a3b8fe-6911-4d8f-bd3e-8044c47a541b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.167860] env[61970]: DEBUG nova.compute.manager [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 625.168080] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 625.168392] env[61970]: DEBUG oslo_concurrency.lockutils [req-8efb9581-5d73-4939-aa5d-a1edce17265f req-53b0870f-0920-4b9a-8673-e87f941dc338 service nova] Acquired lock "refresh_cache-a6a3b8fe-6911-4d8f-bd3e-8044c47a541b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.168600] env[61970]: DEBUG nova.network.neutron [req-8efb9581-5d73-4939-aa5d-a1edce17265f req-53b0870f-0920-4b9a-8673-e87f941dc338 service nova] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Refreshing network info cache for port f3160862-d364-49dd-8639-282b45d418b5 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 625.173392] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1687f355-4feb-42ae-a565-96a8ceb2fba9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.183339] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8463a97d-967b-4ed3-ba0c-cc680d14ccbd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.209698] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a6a3b8fe-6911-4d8f-bd3e-8044c47a541b could not be found. [ 625.209698] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 625.209698] env[61970]: INFO nova.compute.manager [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 625.209698] env[61970]: DEBUG oslo.service.loopingcall [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 625.209698] env[61970]: DEBUG nova.compute.manager [-] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.209698] env[61970]: DEBUG nova.network.neutron [-] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 625.227773] env[61970]: DEBUG nova.network.neutron [-] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.363124] env[61970]: DEBUG nova.network.neutron [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.643216] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "29ac968a-37db-4c0d-9be9-9f577b53c533" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.643470] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "29ac968a-37db-4c0d-9be9-9f577b53c533" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.648728] env[61970]: DEBUG nova.scheduler.client.report [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 625.694516] env[61970]: DEBUG nova.network.neutron [req-8efb9581-5d73-4939-aa5d-a1edce17265f req-53b0870f-0920-4b9a-8673-e87f941dc338 service nova] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.729891] env[61970]: DEBUG nova.network.neutron [-] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.785642] env[61970]: DEBUG nova.network.neutron [req-8efb9581-5d73-4939-aa5d-a1edce17265f req-53b0870f-0920-4b9a-8673-e87f941dc338 service nova] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.867539] env[61970]: INFO nova.compute.manager [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 44ea1800-005c-485f-969d-4bf71d58c284] Took 1.03 seconds to deallocate network for instance. [ 626.153168] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.477s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.155944] env[61970]: DEBUG nova.compute.manager [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 626.160360] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.307s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.163277] env[61970]: INFO nova.compute.claims [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 626.233295] env[61970]: INFO nova.compute.manager [-] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Took 1.02 seconds to deallocate network for instance. [ 626.236384] env[61970]: DEBUG nova.compute.claims [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 626.236621] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.288750] env[61970]: DEBUG oslo_concurrency.lockutils [req-8efb9581-5d73-4939-aa5d-a1edce17265f req-53b0870f-0920-4b9a-8673-e87f941dc338 service nova] Releasing lock "refresh_cache-a6a3b8fe-6911-4d8f-bd3e-8044c47a541b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.290220] env[61970]: DEBUG nova.compute.manager [req-8efb9581-5d73-4939-aa5d-a1edce17265f req-53b0870f-0920-4b9a-8673-e87f941dc338 service nova] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Received event network-vif-deleted-f3160862-d364-49dd-8639-282b45d418b5 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 626.664234] env[61970]: DEBUG nova.compute.utils [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 626.664234] env[61970]: DEBUG nova.compute.manager [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 626.664234] env[61970]: DEBUG nova.network.neutron [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 626.765118] env[61970]: DEBUG nova.policy [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8b9f034f7f474448bfd555c4e8f88f48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '375eb5f90b7141e5a34ab9d79b7f93d5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 626.912812] env[61970]: INFO nova.scheduler.client.report [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Deleted allocations for instance 44ea1800-005c-485f-969d-4bf71d58c284 [ 627.168631] env[61970]: DEBUG nova.compute.manager [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 627.241754] env[61970]: DEBUG nova.network.neutron [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Successfully created port: f82e6809-c983-45f2-b9bf-02b19bfb6043 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 627.430618] env[61970]: DEBUG oslo_concurrency.lockutils [None req-640f73c8-539c-4aa3-913c-c2e05bec2ac1 tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Lock "44ea1800-005c-485f-969d-4bf71d58c284" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.942s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.536347] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2e70f4-4252-4ae8-9a3f-8502215db824 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.545281] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b66006-32d3-4991-8330-34c6b7dfbbaa {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.193915] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 628.193915] env[61970]: DEBUG nova.compute.manager [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 628.193915] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 628.226856] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76426f3e-cb53-459b-a644-2332144f7a36 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.235453] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9672d6e-0610-4bfd-8e75-34f8f156f317 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.251144] env[61970]: DEBUG nova.compute.provider_tree [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.694263] env[61970]: DEBUG nova.compute.manager [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 628.713489] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 628.713789] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Starting heal instance info cache {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 628.713789] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Rebuilding the list of instances to heal {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 628.736185] env[61970]: DEBUG nova.virt.hardware [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 628.736457] env[61970]: DEBUG nova.virt.hardware [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 628.736625] env[61970]: DEBUG nova.virt.hardware [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 628.736755] env[61970]: DEBUG nova.virt.hardware [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 628.736895] env[61970]: DEBUG nova.virt.hardware [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 628.737063] env[61970]: DEBUG nova.virt.hardware [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 628.737282] env[61970]: DEBUG nova.virt.hardware [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 628.737459] env[61970]: DEBUG nova.virt.hardware [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 628.737674] env[61970]: DEBUG nova.virt.hardware [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 628.737803] env[61970]: DEBUG nova.virt.hardware [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 628.737968] env[61970]: DEBUG nova.virt.hardware [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 628.739155] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f451feee-e995-4d3b-b965-ca747dd99513 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.743709] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.750010] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df6ed658-23ec-4416-9f27-f91b6ed91a46 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.754681] env[61970]: DEBUG nova.scheduler.client.report [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.924276] env[61970]: DEBUG nova.compute.manager [req-4bc300de-6766-4fad-b0fc-8f9c88e78d3a req-cc6144f5-cb95-42e4-a997-44a078ae1f93 service nova] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Received event network-changed-f82e6809-c983-45f2-b9bf-02b19bfb6043 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 628.924276] env[61970]: DEBUG nova.compute.manager [req-4bc300de-6766-4fad-b0fc-8f9c88e78d3a req-cc6144f5-cb95-42e4-a997-44a078ae1f93 service nova] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Refreshing instance network info cache due to event network-changed-f82e6809-c983-45f2-b9bf-02b19bfb6043. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 628.924276] env[61970]: DEBUG oslo_concurrency.lockutils [req-4bc300de-6766-4fad-b0fc-8f9c88e78d3a req-cc6144f5-cb95-42e4-a997-44a078ae1f93 service nova] Acquiring lock "refresh_cache-65dba389-321c-478d-9555-bcd9c318f6d0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.924276] env[61970]: DEBUG oslo_concurrency.lockutils [req-4bc300de-6766-4fad-b0fc-8f9c88e78d3a req-cc6144f5-cb95-42e4-a997-44a078ae1f93 service nova] Acquired lock "refresh_cache-65dba389-321c-478d-9555-bcd9c318f6d0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.924276] env[61970]: DEBUG nova.network.neutron [req-4bc300de-6766-4fad-b0fc-8f9c88e78d3a req-cc6144f5-cb95-42e4-a997-44a078ae1f93 service nova] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Refreshing network info cache for port f82e6809-c983-45f2-b9bf-02b19bfb6043 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 629.181362] env[61970]: ERROR nova.compute.manager [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f82e6809-c983-45f2-b9bf-02b19bfb6043, please check neutron logs for more information. [ 629.181362] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 629.181362] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.181362] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 629.181362] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.181362] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 629.181362] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.181362] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 629.181362] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.181362] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 629.181362] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.181362] env[61970]: ERROR nova.compute.manager raise self.value [ 629.181362] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.181362] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 629.181362] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.181362] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 629.181760] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.181760] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 629.181760] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f82e6809-c983-45f2-b9bf-02b19bfb6043, please check neutron logs for more information. [ 629.181760] env[61970]: ERROR nova.compute.manager [ 629.181760] env[61970]: Traceback (most recent call last): [ 629.181760] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 629.181760] env[61970]: listener.cb(fileno) [ 629.181760] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.181760] env[61970]: result = function(*args, **kwargs) [ 629.181760] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 629.181760] env[61970]: return func(*args, **kwargs) [ 629.181760] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.181760] env[61970]: raise e [ 629.181760] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.181760] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 629.181760] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.181760] env[61970]: created_port_ids = self._update_ports_for_instance( [ 629.181760] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.181760] env[61970]: with excutils.save_and_reraise_exception(): [ 629.181760] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.181760] env[61970]: self.force_reraise() [ 629.181760] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.181760] env[61970]: raise self.value [ 629.181760] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.181760] env[61970]: updated_port = self._update_port( [ 629.181760] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.181760] env[61970]: _ensure_no_port_binding_failure(port) [ 629.181760] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.181760] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 629.182845] env[61970]: nova.exception.PortBindingFailed: Binding failed for port f82e6809-c983-45f2-b9bf-02b19bfb6043, please check neutron logs for more information. [ 629.182845] env[61970]: Removing descriptor: 15 [ 629.182845] env[61970]: ERROR nova.compute.manager [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f82e6809-c983-45f2-b9bf-02b19bfb6043, please check neutron logs for more information. [ 629.182845] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Traceback (most recent call last): [ 629.182845] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 629.182845] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] yield resources [ 629.182845] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 629.182845] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] self.driver.spawn(context, instance, image_meta, [ 629.182845] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 629.182845] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.182845] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.182845] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] vm_ref = self.build_virtual_machine(instance, [ 629.183135] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.183135] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.183135] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.183135] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] for vif in network_info: [ 629.183135] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 629.183135] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] return self._sync_wrapper(fn, *args, **kwargs) [ 629.183135] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 629.183135] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] self.wait() [ 629.183135] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 629.183135] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] self[:] = self._gt.wait() [ 629.183135] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.183135] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] return self._exit_event.wait() [ 629.183135] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 629.183435] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] result = hub.switch() [ 629.183435] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 629.183435] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] return self.greenlet.switch() [ 629.183435] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.183435] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] result = function(*args, **kwargs) [ 629.183435] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 629.183435] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] return func(*args, **kwargs) [ 629.183435] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.183435] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] raise e [ 629.183435] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.183435] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] nwinfo = self.network_api.allocate_for_instance( [ 629.183435] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.183435] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] created_port_ids = self._update_ports_for_instance( [ 629.183773] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.183773] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] with excutils.save_and_reraise_exception(): [ 629.183773] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.183773] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] self.force_reraise() [ 629.183773] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.183773] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] raise self.value [ 629.183773] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.183773] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] updated_port = self._update_port( [ 629.183773] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.183773] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] _ensure_no_port_binding_failure(port) [ 629.183773] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.183773] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] raise exception.PortBindingFailed(port_id=port['id']) [ 629.184184] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] nova.exception.PortBindingFailed: Binding failed for port f82e6809-c983-45f2-b9bf-02b19bfb6043, please check neutron logs for more information. [ 629.184184] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] [ 629.184184] env[61970]: INFO nova.compute.manager [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Terminating instance [ 629.188230] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Acquiring lock "refresh_cache-65dba389-321c-478d-9555-bcd9c318f6d0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.219597] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 629.219845] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 629.219897] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 629.220040] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 629.220142] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 629.220253] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Didn't find any instances for network info cache update. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 629.220476] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 629.220587] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 629.220724] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 629.220858] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 629.220990] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 629.221140] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 629.222990] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61970) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 629.223212] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 629.263376] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.101s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.263376] env[61970]: DEBUG nova.compute.manager [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 629.269179] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.219s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.464741] env[61970]: DEBUG nova.network.neutron [req-4bc300de-6766-4fad-b0fc-8f9c88e78d3a req-cc6144f5-cb95-42e4-a997-44a078ae1f93 service nova] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.630116] env[61970]: DEBUG nova.network.neutron [req-4bc300de-6766-4fad-b0fc-8f9c88e78d3a req-cc6144f5-cb95-42e4-a997-44a078ae1f93 service nova] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.729187] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.776708] env[61970]: DEBUG nova.compute.utils [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 629.781335] env[61970]: DEBUG nova.compute.manager [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 629.781335] env[61970]: DEBUG nova.network.neutron [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 629.874629] env[61970]: DEBUG nova.policy [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9c298549c4a948dc85d0e1712c6a9081', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '94b33c82a8514050a75a1257b633d6da', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 630.133072] env[61970]: DEBUG oslo_concurrency.lockutils [req-4bc300de-6766-4fad-b0fc-8f9c88e78d3a req-cc6144f5-cb95-42e4-a997-44a078ae1f93 service nova] Releasing lock "refresh_cache-65dba389-321c-478d-9555-bcd9c318f6d0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.133548] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Acquired lock "refresh_cache-65dba389-321c-478d-9555-bcd9c318f6d0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.134475] env[61970]: DEBUG nova.network.neutron [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 630.192476] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d71cf48-2416-40a2-b592-fc5ad6cc2af1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.200391] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3377c193-5d8f-41e1-a483-a0ae440f4c1b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.234679] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8332cd89-1469-4db0-91c4-f12871af6ae0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.244592] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b514ca0d-28f7-4155-b366-5b0c239564df {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.257442] env[61970]: DEBUG nova.compute.provider_tree [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.279460] env[61970]: DEBUG nova.compute.manager [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 630.388791] env[61970]: DEBUG nova.network.neutron [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Successfully created port: 73c584a8-7f1c-4abd-8b28-60519162a678 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 630.596913] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Acquiring lock "34719fba-dc16-4b43-ab48-517982ca85a4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.596913] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Lock "34719fba-dc16-4b43-ab48-517982ca85a4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.659848] env[61970]: DEBUG nova.network.neutron [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.760308] env[61970]: DEBUG nova.scheduler.client.report [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 630.808921] env[61970]: DEBUG nova.network.neutron [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.948190] env[61970]: DEBUG nova.compute.manager [req-a1585a47-0f5c-41c1-9248-78f2fbc77641 req-6733744e-2569-4315-811c-ead5aae3414a service nova] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Received event network-vif-deleted-f82e6809-c983-45f2-b9bf-02b19bfb6043 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 631.266439] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.998s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.270166] env[61970]: ERROR nova.compute.manager [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f763ac57-bd20-4f40-b28b-d90496239be4, please check neutron logs for more information. [ 631.270166] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Traceback (most recent call last): [ 631.270166] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.270166] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] self.driver.spawn(context, instance, image_meta, [ 631.270166] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 631.270166] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.270166] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.270166] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] vm_ref = self.build_virtual_machine(instance, [ 631.270166] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.270166] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.270166] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.270559] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] for vif in network_info: [ 631.270559] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 631.270559] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] return self._sync_wrapper(fn, *args, **kwargs) [ 631.270559] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 631.270559] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] self.wait() [ 631.270559] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 631.270559] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] self[:] = self._gt.wait() [ 631.270559] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.270559] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] return self._exit_event.wait() [ 631.270559] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 631.270559] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] current.throw(*self._exc) [ 631.270559] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.270559] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] result = function(*args, **kwargs) [ 631.270957] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.270957] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] return func(*args, **kwargs) [ 631.270957] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.270957] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] raise e [ 631.270957] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.270957] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] nwinfo = self.network_api.allocate_for_instance( [ 631.270957] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.270957] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] created_port_ids = self._update_ports_for_instance( [ 631.270957] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.270957] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] with excutils.save_and_reraise_exception(): [ 631.270957] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.270957] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] self.force_reraise() [ 631.270957] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.271255] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] raise self.value [ 631.271255] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.271255] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] updated_port = self._update_port( [ 631.271255] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.271255] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] _ensure_no_port_binding_failure(port) [ 631.271255] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.271255] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] raise exception.PortBindingFailed(port_id=port['id']) [ 631.271255] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] nova.exception.PortBindingFailed: Binding failed for port f763ac57-bd20-4f40-b28b-d90496239be4, please check neutron logs for more information. [ 631.271255] env[61970]: ERROR nova.compute.manager [instance: 59d2c196-9230-4303-9b99-92211c53cb15] [ 631.271255] env[61970]: DEBUG nova.compute.utils [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Binding failed for port f763ac57-bd20-4f40-b28b-d90496239be4, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 631.271489] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.830s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.271489] env[61970]: INFO nova.compute.claims [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 631.273988] env[61970]: DEBUG nova.compute.manager [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Build of instance 59d2c196-9230-4303-9b99-92211c53cb15 was re-scheduled: Binding failed for port f763ac57-bd20-4f40-b28b-d90496239be4, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 631.275237] env[61970]: DEBUG nova.compute.manager [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 631.275237] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Acquiring lock "refresh_cache-59d2c196-9230-4303-9b99-92211c53cb15" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.275237] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Acquired lock "refresh_cache-59d2c196-9230-4303-9b99-92211c53cb15" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.275237] env[61970]: DEBUG nova.network.neutron [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.288303] env[61970]: DEBUG nova.compute.manager [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 631.312455] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Releasing lock "refresh_cache-65dba389-321c-478d-9555-bcd9c318f6d0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.312455] env[61970]: DEBUG nova.compute.manager [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 631.312692] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 631.312912] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9a2a7a76-c6f0-4d74-9174-9a75da581f41 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.317998] env[61970]: DEBUG nova.virt.hardware [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 631.318514] env[61970]: DEBUG nova.virt.hardware [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 631.319097] env[61970]: DEBUG nova.virt.hardware [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 631.319097] env[61970]: DEBUG nova.virt.hardware [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 631.319547] env[61970]: DEBUG nova.virt.hardware [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 631.323201] env[61970]: DEBUG nova.virt.hardware [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 631.323201] env[61970]: DEBUG nova.virt.hardware [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 631.323201] env[61970]: DEBUG nova.virt.hardware [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 631.323201] env[61970]: DEBUG nova.virt.hardware [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 631.323201] env[61970]: DEBUG nova.virt.hardware [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 631.323440] env[61970]: DEBUG nova.virt.hardware [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 631.323440] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361fa5f0-2392-4784-bba6-70f88dd24232 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.335208] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88f560a9-7c0e-4ea9-840d-73d0dc2d0f52 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.344080] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63745ff2-4ae2-4800-a43e-ad996c2de675 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.368768] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 65dba389-321c-478d-9555-bcd9c318f6d0 could not be found. [ 631.369067] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 631.369315] env[61970]: INFO nova.compute.manager [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Took 0.06 seconds to destroy the instance on the hypervisor. [ 631.369597] env[61970]: DEBUG oslo.service.loopingcall [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 631.369904] env[61970]: DEBUG nova.compute.manager [-] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.370067] env[61970]: DEBUG nova.network.neutron [-] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 631.389088] env[61970]: DEBUG nova.network.neutron [-] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.561070] env[61970]: ERROR nova.compute.manager [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 73c584a8-7f1c-4abd-8b28-60519162a678, please check neutron logs for more information. [ 631.561070] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 631.561070] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.561070] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 631.561070] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.561070] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 631.561070] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.561070] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 631.561070] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.561070] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 631.561070] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.561070] env[61970]: ERROR nova.compute.manager raise self.value [ 631.561070] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.561070] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 631.561070] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.561070] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 631.561505] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.561505] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 631.561505] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 73c584a8-7f1c-4abd-8b28-60519162a678, please check neutron logs for more information. [ 631.561505] env[61970]: ERROR nova.compute.manager [ 631.561505] env[61970]: Traceback (most recent call last): [ 631.561505] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 631.561505] env[61970]: listener.cb(fileno) [ 631.561505] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.561505] env[61970]: result = function(*args, **kwargs) [ 631.561505] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.561505] env[61970]: return func(*args, **kwargs) [ 631.561505] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.561505] env[61970]: raise e [ 631.561505] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.561505] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 631.561505] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.561505] env[61970]: created_port_ids = self._update_ports_for_instance( [ 631.561505] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.561505] env[61970]: with excutils.save_and_reraise_exception(): [ 631.561505] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.561505] env[61970]: self.force_reraise() [ 631.561505] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.561505] env[61970]: raise self.value [ 631.561505] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.561505] env[61970]: updated_port = self._update_port( [ 631.561505] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.561505] env[61970]: _ensure_no_port_binding_failure(port) [ 631.561505] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.561505] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 631.562125] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 73c584a8-7f1c-4abd-8b28-60519162a678, please check neutron logs for more information. [ 631.562125] env[61970]: Removing descriptor: 15 [ 631.562125] env[61970]: ERROR nova.compute.manager [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 73c584a8-7f1c-4abd-8b28-60519162a678, please check neutron logs for more information. [ 631.562125] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Traceback (most recent call last): [ 631.562125] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 631.562125] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] yield resources [ 631.562125] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.562125] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] self.driver.spawn(context, instance, image_meta, [ 631.562125] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 631.562125] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.562125] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.562125] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] vm_ref = self.build_virtual_machine(instance, [ 631.562499] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.562499] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.562499] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.562499] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] for vif in network_info: [ 631.562499] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 631.562499] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] return self._sync_wrapper(fn, *args, **kwargs) [ 631.562499] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 631.562499] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] self.wait() [ 631.562499] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 631.562499] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] self[:] = self._gt.wait() [ 631.562499] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.562499] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] return self._exit_event.wait() [ 631.562499] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.562819] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] result = hub.switch() [ 631.562819] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.562819] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] return self.greenlet.switch() [ 631.562819] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.562819] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] result = function(*args, **kwargs) [ 631.562819] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.562819] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] return func(*args, **kwargs) [ 631.562819] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.562819] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] raise e [ 631.562819] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.562819] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] nwinfo = self.network_api.allocate_for_instance( [ 631.562819] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.562819] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] created_port_ids = self._update_ports_for_instance( [ 631.563124] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.563124] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] with excutils.save_and_reraise_exception(): [ 631.563124] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.563124] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] self.force_reraise() [ 631.563124] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.563124] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] raise self.value [ 631.563124] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.563124] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] updated_port = self._update_port( [ 631.563124] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.563124] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] _ensure_no_port_binding_failure(port) [ 631.563124] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.563124] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] raise exception.PortBindingFailed(port_id=port['id']) [ 631.563440] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] nova.exception.PortBindingFailed: Binding failed for port 73c584a8-7f1c-4abd-8b28-60519162a678, please check neutron logs for more information. [ 631.563440] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] [ 631.563440] env[61970]: INFO nova.compute.manager [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Terminating instance [ 631.565633] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Acquiring lock "refresh_cache-0c45c5c0-f8db-4631-ba9a-271d8f9ddff4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.565633] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Acquired lock "refresh_cache-0c45c5c0-f8db-4631-ba9a-271d8f9ddff4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.565633] env[61970]: DEBUG nova.network.neutron [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.720295] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Acquiring lock "1dc83f19-2965-418b-8d2a-56aa1124d258" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.720519] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Lock "1dc83f19-2965-418b-8d2a-56aa1124d258" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.795159] env[61970]: DEBUG nova.network.neutron [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.849098] env[61970]: DEBUG nova.network.neutron [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.891558] env[61970]: DEBUG nova.network.neutron [-] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.083756] env[61970]: DEBUG nova.network.neutron [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.167197] env[61970]: DEBUG nova.network.neutron [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.352122] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Releasing lock "refresh_cache-59d2c196-9230-4303-9b99-92211c53cb15" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.352357] env[61970]: DEBUG nova.compute.manager [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 632.352490] env[61970]: DEBUG nova.compute.manager [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.352657] env[61970]: DEBUG nova.network.neutron [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.375360] env[61970]: DEBUG nova.network.neutron [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.396326] env[61970]: INFO nova.compute.manager [-] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Took 1.02 seconds to deallocate network for instance. [ 632.397090] env[61970]: DEBUG nova.compute.claims [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 632.397761] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.626019] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5179f07-96eb-4617-af09-b7236a212682 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.632704] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d87e3ff-1392-4afd-a3e7-7b67bf63e262 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.665326] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b5c2f17-4302-4f92-9b84-3c8f8503da8d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.669498] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Releasing lock "refresh_cache-0c45c5c0-f8db-4631-ba9a-271d8f9ddff4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.669884] env[61970]: DEBUG nova.compute.manager [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 632.670083] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 632.672724] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f274a20e-24cb-492f-bf14-38996afb9ed3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.675129] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af94bd84-1e6b-48d7-84e2-abe80ee8a055 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.687929] env[61970]: DEBUG nova.compute.provider_tree [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.692816] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f938173-005a-423d-899e-61921dfcd004 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.714951] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4 could not be found. [ 632.715204] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 632.715403] env[61970]: INFO nova.compute.manager [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Took 0.05 seconds to destroy the instance on the hypervisor. [ 632.715672] env[61970]: DEBUG oslo.service.loopingcall [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 632.716436] env[61970]: DEBUG nova.compute.manager [-] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.716436] env[61970]: DEBUG nova.network.neutron [-] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.729597] env[61970]: DEBUG nova.network.neutron [-] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.878571] env[61970]: DEBUG nova.network.neutron [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.974291] env[61970]: DEBUG nova.compute.manager [req-68179aba-accf-441c-ac4d-6b4a7b882610 req-bca16fe7-68cc-4be6-9784-62c05ab11984 service nova] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Received event network-changed-73c584a8-7f1c-4abd-8b28-60519162a678 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 632.974483] env[61970]: DEBUG nova.compute.manager [req-68179aba-accf-441c-ac4d-6b4a7b882610 req-bca16fe7-68cc-4be6-9784-62c05ab11984 service nova] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Refreshing instance network info cache due to event network-changed-73c584a8-7f1c-4abd-8b28-60519162a678. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 632.974697] env[61970]: DEBUG oslo_concurrency.lockutils [req-68179aba-accf-441c-ac4d-6b4a7b882610 req-bca16fe7-68cc-4be6-9784-62c05ab11984 service nova] Acquiring lock "refresh_cache-0c45c5c0-f8db-4631-ba9a-271d8f9ddff4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.974835] env[61970]: DEBUG oslo_concurrency.lockutils [req-68179aba-accf-441c-ac4d-6b4a7b882610 req-bca16fe7-68cc-4be6-9784-62c05ab11984 service nova] Acquired lock "refresh_cache-0c45c5c0-f8db-4631-ba9a-271d8f9ddff4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.974988] env[61970]: DEBUG nova.network.neutron [req-68179aba-accf-441c-ac4d-6b4a7b882610 req-bca16fe7-68cc-4be6-9784-62c05ab11984 service nova] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Refreshing network info cache for port 73c584a8-7f1c-4abd-8b28-60519162a678 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 633.192133] env[61970]: DEBUG nova.scheduler.client.report [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 633.231763] env[61970]: DEBUG nova.network.neutron [-] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.381725] env[61970]: INFO nova.compute.manager [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] [instance: 59d2c196-9230-4303-9b99-92211c53cb15] Took 1.03 seconds to deallocate network for instance. [ 633.499391] env[61970]: DEBUG nova.network.neutron [req-68179aba-accf-441c-ac4d-6b4a7b882610 req-bca16fe7-68cc-4be6-9784-62c05ab11984 service nova] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.585046] env[61970]: DEBUG nova.network.neutron [req-68179aba-accf-441c-ac4d-6b4a7b882610 req-bca16fe7-68cc-4be6-9784-62c05ab11984 service nova] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.699288] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.428s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.699288] env[61970]: DEBUG nova.compute.manager [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 633.700847] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.361s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.702286] env[61970]: INFO nova.compute.claims [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 633.734251] env[61970]: INFO nova.compute.manager [-] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Took 1.02 seconds to deallocate network for instance. [ 633.739182] env[61970]: DEBUG nova.compute.claims [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 633.740497] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.087777] env[61970]: DEBUG oslo_concurrency.lockutils [req-68179aba-accf-441c-ac4d-6b4a7b882610 req-bca16fe7-68cc-4be6-9784-62c05ab11984 service nova] Releasing lock "refresh_cache-0c45c5c0-f8db-4631-ba9a-271d8f9ddff4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.088062] env[61970]: DEBUG nova.compute.manager [req-68179aba-accf-441c-ac4d-6b4a7b882610 req-bca16fe7-68cc-4be6-9784-62c05ab11984 service nova] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Received event network-vif-deleted-73c584a8-7f1c-4abd-8b28-60519162a678 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 634.206708] env[61970]: DEBUG nova.compute.utils [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 634.209905] env[61970]: DEBUG nova.compute.manager [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 634.210091] env[61970]: DEBUG nova.network.neutron [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 634.250525] env[61970]: DEBUG nova.policy [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fbd50319b75e40279be1e2082a7ecb4c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '740d49b85c2f43d3a4a19d20df59da70', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 634.422732] env[61970]: INFO nova.scheduler.client.report [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Deleted allocations for instance 59d2c196-9230-4303-9b99-92211c53cb15 [ 634.650357] env[61970]: DEBUG nova.network.neutron [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Successfully created port: a664bf5a-c723-43e0-a1c2-f19168e6054e {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 634.713442] env[61970]: DEBUG nova.compute.manager [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 634.934354] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2c7a6f5c-3471-46f9-ac96-836a9ffd66ae tempest-DeleteServersAdminTestJSON-1557480121 tempest-DeleteServersAdminTestJSON-1557480121-project-member] Lock "59d2c196-9230-4303-9b99-92211c53cb15" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.173s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.150611] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd6b520b-e22f-46dd-b037-513626a77f08 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.160197] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a577eeb9-0b88-4b5f-917c-17b22c6e1b97 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.191287] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70fa24dd-a3c4-4ac5-975c-688d7b7c78ba {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.198731] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7bdd602-ad36-47d5-8b9c-20fb48f77199 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.211930] env[61970]: DEBUG nova.compute.provider_tree [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.442072] env[61970]: DEBUG nova.compute.manager [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 635.717517] env[61970]: DEBUG nova.scheduler.client.report [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 635.728958] env[61970]: DEBUG nova.compute.manager [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 635.758446] env[61970]: DEBUG nova.virt.hardware [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 635.758728] env[61970]: DEBUG nova.virt.hardware [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 635.758884] env[61970]: DEBUG nova.virt.hardware [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 635.759074] env[61970]: DEBUG nova.virt.hardware [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 635.759219] env[61970]: DEBUG nova.virt.hardware [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 635.759364] env[61970]: DEBUG nova.virt.hardware [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 635.759571] env[61970]: DEBUG nova.virt.hardware [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 635.759731] env[61970]: DEBUG nova.virt.hardware [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 635.759894] env[61970]: DEBUG nova.virt.hardware [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 635.764224] env[61970]: DEBUG nova.virt.hardware [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 635.764459] env[61970]: DEBUG nova.virt.hardware [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 635.765368] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c09cf33e-abff-49c5-82eb-c881b641e780 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.777533] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-928ca1c9-93b3-4573-961e-9b14a138e693 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.850516] env[61970]: DEBUG nova.compute.manager [req-78ee346a-4820-4e2e-b79f-785e6ce827ee req-efe33a74-0d8b-49e1-979b-45ecc3f52326 service nova] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Received event network-changed-a664bf5a-c723-43e0-a1c2-f19168e6054e {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 635.850516] env[61970]: DEBUG nova.compute.manager [req-78ee346a-4820-4e2e-b79f-785e6ce827ee req-efe33a74-0d8b-49e1-979b-45ecc3f52326 service nova] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Refreshing instance network info cache due to event network-changed-a664bf5a-c723-43e0-a1c2-f19168e6054e. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 635.850720] env[61970]: DEBUG oslo_concurrency.lockutils [req-78ee346a-4820-4e2e-b79f-785e6ce827ee req-efe33a74-0d8b-49e1-979b-45ecc3f52326 service nova] Acquiring lock "refresh_cache-816de9f9-a8a9-415a-a6c4-57baea86df51" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.850720] env[61970]: DEBUG oslo_concurrency.lockutils [req-78ee346a-4820-4e2e-b79f-785e6ce827ee req-efe33a74-0d8b-49e1-979b-45ecc3f52326 service nova] Acquired lock "refresh_cache-816de9f9-a8a9-415a-a6c4-57baea86df51" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.850875] env[61970]: DEBUG nova.network.neutron [req-78ee346a-4820-4e2e-b79f-785e6ce827ee req-efe33a74-0d8b-49e1-979b-45ecc3f52326 service nova] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Refreshing network info cache for port a664bf5a-c723-43e0-a1c2-f19168e6054e {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 635.878841] env[61970]: ERROR nova.compute.manager [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a664bf5a-c723-43e0-a1c2-f19168e6054e, please check neutron logs for more information. [ 635.878841] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 635.878841] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.878841] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 635.878841] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.878841] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 635.878841] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.878841] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 635.878841] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.878841] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 635.878841] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.878841] env[61970]: ERROR nova.compute.manager raise self.value [ 635.878841] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.878841] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 635.878841] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.878841] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 635.879343] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.879343] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 635.879343] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a664bf5a-c723-43e0-a1c2-f19168e6054e, please check neutron logs for more information. [ 635.879343] env[61970]: ERROR nova.compute.manager [ 635.879343] env[61970]: Traceback (most recent call last): [ 635.879343] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 635.879343] env[61970]: listener.cb(fileno) [ 635.879343] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.879343] env[61970]: result = function(*args, **kwargs) [ 635.879343] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 635.879343] env[61970]: return func(*args, **kwargs) [ 635.879343] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.879343] env[61970]: raise e [ 635.879343] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.879343] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 635.879343] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.879343] env[61970]: created_port_ids = self._update_ports_for_instance( [ 635.879343] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.879343] env[61970]: with excutils.save_and_reraise_exception(): [ 635.879343] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.879343] env[61970]: self.force_reraise() [ 635.879343] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.879343] env[61970]: raise self.value [ 635.879343] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.879343] env[61970]: updated_port = self._update_port( [ 635.879343] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.879343] env[61970]: _ensure_no_port_binding_failure(port) [ 635.879343] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.879343] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 635.880136] env[61970]: nova.exception.PortBindingFailed: Binding failed for port a664bf5a-c723-43e0-a1c2-f19168e6054e, please check neutron logs for more information. [ 635.880136] env[61970]: Removing descriptor: 15 [ 635.880136] env[61970]: ERROR nova.compute.manager [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a664bf5a-c723-43e0-a1c2-f19168e6054e, please check neutron logs for more information. [ 635.880136] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Traceback (most recent call last): [ 635.880136] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 635.880136] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] yield resources [ 635.880136] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 635.880136] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] self.driver.spawn(context, instance, image_meta, [ 635.880136] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 635.880136] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.880136] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.880136] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] vm_ref = self.build_virtual_machine(instance, [ 635.880435] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.880435] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.880435] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.880435] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] for vif in network_info: [ 635.880435] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 635.880435] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] return self._sync_wrapper(fn, *args, **kwargs) [ 635.880435] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 635.880435] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] self.wait() [ 635.880435] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 635.880435] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] self[:] = self._gt.wait() [ 635.880435] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.880435] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] return self._exit_event.wait() [ 635.880435] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 635.880726] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] result = hub.switch() [ 635.880726] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 635.880726] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] return self.greenlet.switch() [ 635.880726] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.880726] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] result = function(*args, **kwargs) [ 635.880726] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 635.880726] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] return func(*args, **kwargs) [ 635.880726] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.880726] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] raise e [ 635.880726] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.880726] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] nwinfo = self.network_api.allocate_for_instance( [ 635.880726] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.880726] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] created_port_ids = self._update_ports_for_instance( [ 635.881071] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.881071] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] with excutils.save_and_reraise_exception(): [ 635.881071] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.881071] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] self.force_reraise() [ 635.881071] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.881071] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] raise self.value [ 635.881071] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.881071] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] updated_port = self._update_port( [ 635.881071] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.881071] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] _ensure_no_port_binding_failure(port) [ 635.881071] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.881071] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] raise exception.PortBindingFailed(port_id=port['id']) [ 635.881880] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] nova.exception.PortBindingFailed: Binding failed for port a664bf5a-c723-43e0-a1c2-f19168e6054e, please check neutron logs for more information. [ 635.881880] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] [ 635.881880] env[61970]: INFO nova.compute.manager [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Terminating instance [ 635.883017] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Acquiring lock "refresh_cache-816de9f9-a8a9-415a-a6c4-57baea86df51" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.960545] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.223992] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.224550] env[61970]: DEBUG nova.compute.manager [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 636.227320] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.899s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.228733] env[61970]: INFO nova.compute.claims [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 636.372208] env[61970]: DEBUG nova.network.neutron [req-78ee346a-4820-4e2e-b79f-785e6ce827ee req-efe33a74-0d8b-49e1-979b-45ecc3f52326 service nova] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.465502] env[61970]: DEBUG nova.network.neutron [req-78ee346a-4820-4e2e-b79f-785e6ce827ee req-efe33a74-0d8b-49e1-979b-45ecc3f52326 service nova] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.733400] env[61970]: DEBUG nova.compute.utils [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 636.736473] env[61970]: DEBUG nova.compute.manager [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 636.736674] env[61970]: DEBUG nova.network.neutron [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 636.793300] env[61970]: DEBUG nova.policy [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '669cb3cf89814078b1356acc54b85e63', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9670b23779f54b329ad99e573812c0e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 636.968360] env[61970]: DEBUG oslo_concurrency.lockutils [req-78ee346a-4820-4e2e-b79f-785e6ce827ee req-efe33a74-0d8b-49e1-979b-45ecc3f52326 service nova] Releasing lock "refresh_cache-816de9f9-a8a9-415a-a6c4-57baea86df51" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.968861] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Acquired lock "refresh_cache-816de9f9-a8a9-415a-a6c4-57baea86df51" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.969108] env[61970]: DEBUG nova.network.neutron [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 637.242879] env[61970]: DEBUG nova.compute.manager [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 637.291618] env[61970]: DEBUG nova.network.neutron [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Successfully created port: f2005d63-aefe-4dc0-b331-c296a931017f {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 637.492298] env[61970]: DEBUG nova.network.neutron [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.596648] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d91136ee-cc4d-4c09-b172-f61e8df0b689 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.605313] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101e09b1-b440-4208-a1a1-9ac6735deb12 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.609486] env[61970]: DEBUG nova.network.neutron [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.646594] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Releasing lock "refresh_cache-816de9f9-a8a9-415a-a6c4-57baea86df51" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.646979] env[61970]: DEBUG nova.compute.manager [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 637.647178] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 637.647983] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7179a9be-756c-4375-9d61-63ebb655f096 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.650870] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-03140357-cd39-4006-8620-08505a94f870 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.658291] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84b6d08-c87d-4917-ad96-ad5226186568 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.664815] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be73283-ce92-4edc-8014-adf12d5d1404 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.684982] env[61970]: DEBUG nova.compute.provider_tree [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.689932] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 816de9f9-a8a9-415a-a6c4-57baea86df51 could not be found. [ 637.689932] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 637.690104] env[61970]: INFO nova.compute.manager [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Took 0.04 seconds to destroy the instance on the hypervisor. [ 637.690214] env[61970]: DEBUG oslo.service.loopingcall [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 637.690604] env[61970]: DEBUG nova.compute.manager [-] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 637.690699] env[61970]: DEBUG nova.network.neutron [-] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 637.708232] env[61970]: DEBUG nova.network.neutron [-] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.907137] env[61970]: DEBUG nova.compute.manager [req-07aaa42b-da56-41b7-8bc6-5b6d7aa4702f req-6f44ec8e-fac4-46fc-b11e-070246574036 service nova] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Received event network-vif-deleted-a664bf5a-c723-43e0-a1c2-f19168e6054e {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 638.193098] env[61970]: DEBUG nova.scheduler.client.report [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 638.211382] env[61970]: DEBUG nova.network.neutron [-] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.252760] env[61970]: DEBUG nova.compute.manager [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 638.284276] env[61970]: DEBUG nova.virt.hardware [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 638.285031] env[61970]: DEBUG nova.virt.hardware [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 638.285031] env[61970]: DEBUG nova.virt.hardware [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 638.285031] env[61970]: DEBUG nova.virt.hardware [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 638.285031] env[61970]: DEBUG nova.virt.hardware [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 638.285215] env[61970]: DEBUG nova.virt.hardware [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 638.285333] env[61970]: DEBUG nova.virt.hardware [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 638.285542] env[61970]: DEBUG nova.virt.hardware [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 638.285723] env[61970]: DEBUG nova.virt.hardware [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 638.289022] env[61970]: DEBUG nova.virt.hardware [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 638.289022] env[61970]: DEBUG nova.virt.hardware [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 638.289022] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cc345a3-a070-441c-b02b-81a7ebfd50fc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.295970] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f228242-33f9-4c6d-a910-5ae674faaa1b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.385611] env[61970]: ERROR nova.compute.manager [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f2005d63-aefe-4dc0-b331-c296a931017f, please check neutron logs for more information. [ 638.385611] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 638.385611] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.385611] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 638.385611] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.385611] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 638.385611] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.385611] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 638.385611] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.385611] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 638.385611] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.385611] env[61970]: ERROR nova.compute.manager raise self.value [ 638.385611] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.385611] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 638.385611] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.385611] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 638.386179] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.386179] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 638.386179] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f2005d63-aefe-4dc0-b331-c296a931017f, please check neutron logs for more information. [ 638.386179] env[61970]: ERROR nova.compute.manager [ 638.386179] env[61970]: Traceback (most recent call last): [ 638.386179] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 638.386179] env[61970]: listener.cb(fileno) [ 638.386179] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.386179] env[61970]: result = function(*args, **kwargs) [ 638.386179] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 638.386179] env[61970]: return func(*args, **kwargs) [ 638.386179] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.386179] env[61970]: raise e [ 638.386179] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.386179] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 638.386179] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.386179] env[61970]: created_port_ids = self._update_ports_for_instance( [ 638.386179] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.386179] env[61970]: with excutils.save_and_reraise_exception(): [ 638.386179] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.386179] env[61970]: self.force_reraise() [ 638.386179] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.386179] env[61970]: raise self.value [ 638.386179] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.386179] env[61970]: updated_port = self._update_port( [ 638.386179] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.386179] env[61970]: _ensure_no_port_binding_failure(port) [ 638.386179] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.386179] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 638.387273] env[61970]: nova.exception.PortBindingFailed: Binding failed for port f2005d63-aefe-4dc0-b331-c296a931017f, please check neutron logs for more information. [ 638.387273] env[61970]: Removing descriptor: 15 [ 638.387273] env[61970]: ERROR nova.compute.manager [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f2005d63-aefe-4dc0-b331-c296a931017f, please check neutron logs for more information. [ 638.387273] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Traceback (most recent call last): [ 638.387273] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 638.387273] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] yield resources [ 638.387273] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 638.387273] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] self.driver.spawn(context, instance, image_meta, [ 638.387273] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 638.387273] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.387273] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.387273] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] vm_ref = self.build_virtual_machine(instance, [ 638.387670] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.387670] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.387670] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.387670] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] for vif in network_info: [ 638.387670] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 638.387670] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] return self._sync_wrapper(fn, *args, **kwargs) [ 638.387670] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 638.387670] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] self.wait() [ 638.387670] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 638.387670] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] self[:] = self._gt.wait() [ 638.387670] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.387670] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] return self._exit_event.wait() [ 638.387670] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 638.388060] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] result = hub.switch() [ 638.388060] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 638.388060] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] return self.greenlet.switch() [ 638.388060] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.388060] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] result = function(*args, **kwargs) [ 638.388060] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 638.388060] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] return func(*args, **kwargs) [ 638.388060] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.388060] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] raise e [ 638.388060] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.388060] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] nwinfo = self.network_api.allocate_for_instance( [ 638.388060] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.388060] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] created_port_ids = self._update_ports_for_instance( [ 638.388549] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.388549] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] with excutils.save_and_reraise_exception(): [ 638.388549] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.388549] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] self.force_reraise() [ 638.388549] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.388549] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] raise self.value [ 638.388549] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.388549] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] updated_port = self._update_port( [ 638.388549] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.388549] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] _ensure_no_port_binding_failure(port) [ 638.388549] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.388549] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] raise exception.PortBindingFailed(port_id=port['id']) [ 638.388893] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] nova.exception.PortBindingFailed: Binding failed for port f2005d63-aefe-4dc0-b331-c296a931017f, please check neutron logs for more information. [ 638.388893] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] [ 638.388893] env[61970]: INFO nova.compute.manager [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Terminating instance [ 638.389848] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Acquiring lock "refresh_cache-26f6843a-2548-47e0-9dc9-bc53eda32b60" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.390053] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Acquired lock "refresh_cache-26f6843a-2548-47e0-9dc9-bc53eda32b60" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.390259] env[61970]: DEBUG nova.network.neutron [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.703024] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.703024] env[61970]: DEBUG nova.compute.manager [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 638.706462] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.009s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.715930] env[61970]: INFO nova.compute.manager [-] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Took 1.03 seconds to deallocate network for instance. [ 638.719494] env[61970]: DEBUG nova.compute.claims [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 638.719494] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.914820] env[61970]: DEBUG nova.network.neutron [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.024564] env[61970]: DEBUG nova.network.neutron [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.213857] env[61970]: DEBUG nova.compute.utils [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 639.219554] env[61970]: DEBUG nova.compute.manager [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 639.219626] env[61970]: DEBUG nova.network.neutron [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 639.277727] env[61970]: DEBUG nova.policy [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b45dc9d71c0e4693ba42322efde6c572', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5c64cb9232fb413cbd7627dcf077e9ef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 639.529460] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Releasing lock "refresh_cache-26f6843a-2548-47e0-9dc9-bc53eda32b60" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.529878] env[61970]: DEBUG nova.compute.manager [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 639.530105] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 639.530435] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8ea7e29b-d57e-41ff-a9a8-78db947929de {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.543843] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d271f1d-f9d3-4a91-b4a9-8f2edc8173a0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.569179] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 26f6843a-2548-47e0-9dc9-bc53eda32b60 could not be found. [ 639.569422] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 639.569605] env[61970]: INFO nova.compute.manager [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Took 0.04 seconds to destroy the instance on the hypervisor. [ 639.569859] env[61970]: DEBUG oslo.service.loopingcall [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 639.570090] env[61970]: DEBUG nova.compute.manager [-] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.570183] env[61970]: DEBUG nova.network.neutron [-] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.598549] env[61970]: DEBUG nova.network.neutron [-] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.638921] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54d916c2-1dc6-4714-a9bd-336937892a79 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.650389] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13bff264-6b02-4284-b3f0-c30476c8dfa8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.684598] env[61970]: DEBUG nova.network.neutron [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Successfully created port: 379ab7e2-3794-4982-add9-60278acf0764 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 639.686983] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7183a04-75ce-4003-a493-7622ea195acb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.694964] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8aa2db3-58b5-4ce7-8ee1-790d6dd31573 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.710437] env[61970]: DEBUG nova.compute.provider_tree [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.720677] env[61970]: DEBUG nova.compute.manager [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 639.934092] env[61970]: DEBUG nova.compute.manager [req-48560d44-87e5-4739-af79-6924e6dddb6e req-ae6bc493-d855-4cc2-935e-e331ce4e9f32 service nova] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Received event network-changed-f2005d63-aefe-4dc0-b331-c296a931017f {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 639.934311] env[61970]: DEBUG nova.compute.manager [req-48560d44-87e5-4739-af79-6924e6dddb6e req-ae6bc493-d855-4cc2-935e-e331ce4e9f32 service nova] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Refreshing instance network info cache due to event network-changed-f2005d63-aefe-4dc0-b331-c296a931017f. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 639.934527] env[61970]: DEBUG oslo_concurrency.lockutils [req-48560d44-87e5-4739-af79-6924e6dddb6e req-ae6bc493-d855-4cc2-935e-e331ce4e9f32 service nova] Acquiring lock "refresh_cache-26f6843a-2548-47e0-9dc9-bc53eda32b60" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.934669] env[61970]: DEBUG oslo_concurrency.lockutils [req-48560d44-87e5-4739-af79-6924e6dddb6e req-ae6bc493-d855-4cc2-935e-e331ce4e9f32 service nova] Acquired lock "refresh_cache-26f6843a-2548-47e0-9dc9-bc53eda32b60" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.935140] env[61970]: DEBUG nova.network.neutron [req-48560d44-87e5-4739-af79-6924e6dddb6e req-ae6bc493-d855-4cc2-935e-e331ce4e9f32 service nova] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Refreshing network info cache for port f2005d63-aefe-4dc0-b331-c296a931017f {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 640.104296] env[61970]: DEBUG nova.network.neutron [-] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.213727] env[61970]: DEBUG nova.scheduler.client.report [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 640.457099] env[61970]: DEBUG nova.network.neutron [req-48560d44-87e5-4739-af79-6924e6dddb6e req-ae6bc493-d855-4cc2-935e-e331ce4e9f32 service nova] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.581449] env[61970]: DEBUG nova.network.neutron [req-48560d44-87e5-4739-af79-6924e6dddb6e req-ae6bc493-d855-4cc2-935e-e331ce4e9f32 service nova] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.607227] env[61970]: INFO nova.compute.manager [-] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Took 1.04 seconds to deallocate network for instance. [ 640.609975] env[61970]: DEBUG nova.compute.claims [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 640.610174] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.719304] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.012s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.720054] env[61970]: ERROR nova.compute.manager [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d01aba36-d853-4a23-b25f-dedaf27edf79, please check neutron logs for more information. [ 640.720054] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Traceback (most recent call last): [ 640.720054] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 640.720054] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] self.driver.spawn(context, instance, image_meta, [ 640.720054] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 640.720054] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.720054] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.720054] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] vm_ref = self.build_virtual_machine(instance, [ 640.720054] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.720054] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.720054] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.720551] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] for vif in network_info: [ 640.720551] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 640.720551] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] return self._sync_wrapper(fn, *args, **kwargs) [ 640.720551] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 640.720551] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] self.wait() [ 640.720551] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 640.720551] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] self[:] = self._gt.wait() [ 640.720551] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.720551] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] return self._exit_event.wait() [ 640.720551] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 640.720551] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] result = hub.switch() [ 640.720551] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 640.720551] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] return self.greenlet.switch() [ 640.721078] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.721078] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] result = function(*args, **kwargs) [ 640.721078] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 640.721078] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] return func(*args, **kwargs) [ 640.721078] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.721078] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] raise e [ 640.721078] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.721078] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] nwinfo = self.network_api.allocate_for_instance( [ 640.721078] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.721078] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] created_port_ids = self._update_ports_for_instance( [ 640.721078] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.721078] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] with excutils.save_and_reraise_exception(): [ 640.721078] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.721891] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] self.force_reraise() [ 640.721891] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.721891] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] raise self.value [ 640.721891] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.721891] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] updated_port = self._update_port( [ 640.721891] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.721891] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] _ensure_no_port_binding_failure(port) [ 640.721891] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.721891] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] raise exception.PortBindingFailed(port_id=port['id']) [ 640.721891] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] nova.exception.PortBindingFailed: Binding failed for port d01aba36-d853-4a23-b25f-dedaf27edf79, please check neutron logs for more information. [ 640.721891] env[61970]: ERROR nova.compute.manager [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] [ 640.722500] env[61970]: DEBUG nova.compute.utils [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Binding failed for port d01aba36-d853-4a23-b25f-dedaf27edf79, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 640.722500] env[61970]: DEBUG oslo_concurrency.lockutils [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.620s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.727023] env[61970]: INFO nova.compute.claims [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 640.727404] env[61970]: DEBUG nova.compute.manager [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Build of instance 7c48e5b6-eb5f-4c35-9966-1cfc69182595 was re-scheduled: Binding failed for port d01aba36-d853-4a23-b25f-dedaf27edf79, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 640.728150] env[61970]: DEBUG nova.compute.manager [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 640.728369] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "refresh_cache-7c48e5b6-eb5f-4c35-9966-1cfc69182595" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.728527] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquired lock "refresh_cache-7c48e5b6-eb5f-4c35-9966-1cfc69182595" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.729398] env[61970]: DEBUG nova.network.neutron [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 640.733924] env[61970]: DEBUG nova.compute.manager [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 640.761790] env[61970]: DEBUG nova.virt.hardware [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 640.762628] env[61970]: DEBUG nova.virt.hardware [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 640.762835] env[61970]: DEBUG nova.virt.hardware [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 640.763561] env[61970]: DEBUG nova.virt.hardware [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 640.763794] env[61970]: DEBUG nova.virt.hardware [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 640.764135] env[61970]: DEBUG nova.virt.hardware [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 640.765020] env[61970]: DEBUG nova.virt.hardware [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 640.765247] env[61970]: DEBUG nova.virt.hardware [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 640.765435] env[61970]: DEBUG nova.virt.hardware [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 640.765720] env[61970]: DEBUG nova.virt.hardware [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 640.765774] env[61970]: DEBUG nova.virt.hardware [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 640.766777] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51f3ab7c-4467-47a1-89cd-1cf675dbdb91 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.777643] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb538fa8-0b6e-4b98-ad6d-f92461aaee1e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.816904] env[61970]: ERROR nova.compute.manager [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 379ab7e2-3794-4982-add9-60278acf0764, please check neutron logs for more information. [ 640.816904] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 640.816904] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.816904] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 640.816904] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.816904] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 640.816904] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.816904] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 640.816904] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.816904] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 640.816904] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.816904] env[61970]: ERROR nova.compute.manager raise self.value [ 640.816904] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.816904] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 640.816904] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.816904] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 640.817304] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.817304] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 640.817304] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 379ab7e2-3794-4982-add9-60278acf0764, please check neutron logs for more information. [ 640.817304] env[61970]: ERROR nova.compute.manager [ 640.817304] env[61970]: Traceback (most recent call last): [ 640.817304] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 640.817304] env[61970]: listener.cb(fileno) [ 640.817304] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.817304] env[61970]: result = function(*args, **kwargs) [ 640.817304] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 640.817304] env[61970]: return func(*args, **kwargs) [ 640.817304] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.817304] env[61970]: raise e [ 640.817304] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.817304] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 640.817304] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.817304] env[61970]: created_port_ids = self._update_ports_for_instance( [ 640.817304] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.817304] env[61970]: with excutils.save_and_reraise_exception(): [ 640.817304] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.817304] env[61970]: self.force_reraise() [ 640.817304] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.817304] env[61970]: raise self.value [ 640.817304] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.817304] env[61970]: updated_port = self._update_port( [ 640.817304] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.817304] env[61970]: _ensure_no_port_binding_failure(port) [ 640.817304] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.817304] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 640.817971] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 379ab7e2-3794-4982-add9-60278acf0764, please check neutron logs for more information. [ 640.817971] env[61970]: Removing descriptor: 15 [ 640.817971] env[61970]: ERROR nova.compute.manager [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 379ab7e2-3794-4982-add9-60278acf0764, please check neutron logs for more information. [ 640.817971] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Traceback (most recent call last): [ 640.817971] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 640.817971] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] yield resources [ 640.817971] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 640.817971] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] self.driver.spawn(context, instance, image_meta, [ 640.817971] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 640.817971] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.817971] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.817971] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] vm_ref = self.build_virtual_machine(instance, [ 640.818292] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.818292] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.818292] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.818292] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] for vif in network_info: [ 640.818292] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 640.818292] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] return self._sync_wrapper(fn, *args, **kwargs) [ 640.818292] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 640.818292] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] self.wait() [ 640.818292] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 640.818292] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] self[:] = self._gt.wait() [ 640.818292] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.818292] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] return self._exit_event.wait() [ 640.818292] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 640.818618] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] result = hub.switch() [ 640.818618] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 640.818618] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] return self.greenlet.switch() [ 640.818618] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.818618] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] result = function(*args, **kwargs) [ 640.818618] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 640.818618] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] return func(*args, **kwargs) [ 640.818618] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.818618] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] raise e [ 640.818618] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.818618] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] nwinfo = self.network_api.allocate_for_instance( [ 640.818618] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.818618] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] created_port_ids = self._update_ports_for_instance( [ 640.818921] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.818921] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] with excutils.save_and_reraise_exception(): [ 640.818921] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.818921] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] self.force_reraise() [ 640.818921] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.818921] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] raise self.value [ 640.818921] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.818921] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] updated_port = self._update_port( [ 640.818921] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.818921] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] _ensure_no_port_binding_failure(port) [ 640.818921] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.818921] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] raise exception.PortBindingFailed(port_id=port['id']) [ 640.819203] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] nova.exception.PortBindingFailed: Binding failed for port 379ab7e2-3794-4982-add9-60278acf0764, please check neutron logs for more information. [ 640.819203] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] [ 640.819203] env[61970]: INFO nova.compute.manager [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Terminating instance [ 640.821186] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "refresh_cache-2933b91a-ab88-4f0f-bd63-3752cdc763de" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.821404] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "refresh_cache-2933b91a-ab88-4f0f-bd63-3752cdc763de" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.821848] env[61970]: DEBUG nova.network.neutron [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 641.089705] env[61970]: DEBUG oslo_concurrency.lockutils [req-48560d44-87e5-4739-af79-6924e6dddb6e req-ae6bc493-d855-4cc2-935e-e331ce4e9f32 service nova] Releasing lock "refresh_cache-26f6843a-2548-47e0-9dc9-bc53eda32b60" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.089971] env[61970]: DEBUG nova.compute.manager [req-48560d44-87e5-4739-af79-6924e6dddb6e req-ae6bc493-d855-4cc2-935e-e331ce4e9f32 service nova] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Received event network-vif-deleted-f2005d63-aefe-4dc0-b331-c296a931017f {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 641.270043] env[61970]: DEBUG nova.network.neutron [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.354060] env[61970]: DEBUG nova.network.neutron [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.418574] env[61970]: DEBUG nova.network.neutron [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.474549] env[61970]: DEBUG nova.network.neutron [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.921696] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Releasing lock "refresh_cache-7c48e5b6-eb5f-4c35-9966-1cfc69182595" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.922019] env[61970]: DEBUG nova.compute.manager [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 641.922137] env[61970]: DEBUG nova.compute.manager [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 641.922258] env[61970]: DEBUG nova.network.neutron [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 641.944372] env[61970]: DEBUG nova.network.neutron [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.963924] env[61970]: DEBUG nova.compute.manager [req-28f64332-37e1-4c25-af36-e606df60db47 req-518fb8e7-16cd-47d3-875b-345fb314f975 service nova] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Received event network-changed-379ab7e2-3794-4982-add9-60278acf0764 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 641.964126] env[61970]: DEBUG nova.compute.manager [req-28f64332-37e1-4c25-af36-e606df60db47 req-518fb8e7-16cd-47d3-875b-345fb314f975 service nova] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Refreshing instance network info cache due to event network-changed-379ab7e2-3794-4982-add9-60278acf0764. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 641.964761] env[61970]: DEBUG oslo_concurrency.lockutils [req-28f64332-37e1-4c25-af36-e606df60db47 req-518fb8e7-16cd-47d3-875b-345fb314f975 service nova] Acquiring lock "refresh_cache-2933b91a-ab88-4f0f-bd63-3752cdc763de" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.977641] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "refresh_cache-2933b91a-ab88-4f0f-bd63-3752cdc763de" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.977941] env[61970]: DEBUG nova.compute.manager [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 641.978185] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 641.980671] env[61970]: DEBUG oslo_concurrency.lockutils [req-28f64332-37e1-4c25-af36-e606df60db47 req-518fb8e7-16cd-47d3-875b-345fb314f975 service nova] Acquired lock "refresh_cache-2933b91a-ab88-4f0f-bd63-3752cdc763de" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.980924] env[61970]: DEBUG nova.network.neutron [req-28f64332-37e1-4c25-af36-e606df60db47 req-518fb8e7-16cd-47d3-875b-345fb314f975 service nova] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Refreshing network info cache for port 379ab7e2-3794-4982-add9-60278acf0764 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 641.981955] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-55a6c915-e1cf-49fd-a977-f177dc97692f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.991353] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d219caa-71ba-425e-b437-433967560cff {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.016469] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2933b91a-ab88-4f0f-bd63-3752cdc763de could not be found. [ 642.016687] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 642.016917] env[61970]: INFO nova.compute.manager [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Took 0.04 seconds to destroy the instance on the hypervisor. [ 642.017112] env[61970]: DEBUG oslo.service.loopingcall [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 642.019724] env[61970]: DEBUG nova.compute.manager [-] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 642.019785] env[61970]: DEBUG nova.network.neutron [-] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 642.035830] env[61970]: DEBUG nova.network.neutron [-] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.089022] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db785ee7-949c-4fd9-89d2-26acc32e6568 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.098073] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d95da17e-20bd-4ef3-a6a4-ea6b31fb29d8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.130049] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2026686-048b-4c68-a73a-94d8827c33c0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.137727] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6193ce08-5c42-4e16-8cc8-8f67de1ae5fc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.151128] env[61970]: DEBUG nova.compute.provider_tree [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.450831] env[61970]: DEBUG nova.network.neutron [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.502550] env[61970]: DEBUG nova.network.neutron [req-28f64332-37e1-4c25-af36-e606df60db47 req-518fb8e7-16cd-47d3-875b-345fb314f975 service nova] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.541406] env[61970]: DEBUG nova.network.neutron [-] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.604406] env[61970]: DEBUG nova.network.neutron [req-28f64332-37e1-4c25-af36-e606df60db47 req-518fb8e7-16cd-47d3-875b-345fb314f975 service nova] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.655355] env[61970]: DEBUG nova.scheduler.client.report [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 642.799175] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Acquiring lock "ed10e0a6-c79f-4783-bb93-312c5f0ffd51" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.799357] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Lock "ed10e0a6-c79f-4783-bb93-312c5f0ffd51" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.953634] env[61970]: INFO nova.compute.manager [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 7c48e5b6-eb5f-4c35-9966-1cfc69182595] Took 1.03 seconds to deallocate network for instance. [ 643.045917] env[61970]: INFO nova.compute.manager [-] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Took 1.03 seconds to deallocate network for instance. [ 643.048244] env[61970]: DEBUG nova.compute.claims [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 643.048427] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.109054] env[61970]: DEBUG oslo_concurrency.lockutils [req-28f64332-37e1-4c25-af36-e606df60db47 req-518fb8e7-16cd-47d3-875b-345fb314f975 service nova] Releasing lock "refresh_cache-2933b91a-ab88-4f0f-bd63-3752cdc763de" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.109392] env[61970]: DEBUG nova.compute.manager [req-28f64332-37e1-4c25-af36-e606df60db47 req-518fb8e7-16cd-47d3-875b-345fb314f975 service nova] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Received event network-vif-deleted-379ab7e2-3794-4982-add9-60278acf0764 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 643.137256] env[61970]: DEBUG oslo_concurrency.lockutils [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Acquiring lock "5d352f39-4544-46ec-af43-e539d511b47b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.137256] env[61970]: DEBUG oslo_concurrency.lockutils [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Lock "5d352f39-4544-46ec-af43-e539d511b47b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.161024] env[61970]: DEBUG oslo_concurrency.lockutils [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.438s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.161024] env[61970]: DEBUG nova.compute.manager [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 643.163052] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.926s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.672022] env[61970]: DEBUG nova.compute.utils [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 643.675185] env[61970]: DEBUG nova.compute.manager [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 643.678184] env[61970]: DEBUG nova.network.neutron [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 643.752102] env[61970]: DEBUG nova.policy [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ef80257aeedb480ebfca4238d2cd7c8e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91807992e5154986ba4c6a51616397aa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 643.982509] env[61970]: INFO nova.scheduler.client.report [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Deleted allocations for instance 7c48e5b6-eb5f-4c35-9966-1cfc69182595 [ 644.085030] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae5ad95-01b9-4d0c-b83a-aa73492af2d0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.093206] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d41c3d4c-8192-462a-9c58-5d559ba8e0d7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.096976] env[61970]: DEBUG nova.network.neutron [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Successfully created port: 58742262-bf56-4023-8e83-5d4ddd05cfcb {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 644.129769] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089c579e-2187-4e2c-b5ea-93e8c4d5530d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.137621] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e1b37d1-20b9-4343-ae51-2696f4801b66 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.152466] env[61970]: DEBUG nova.compute.provider_tree [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.178459] env[61970]: DEBUG nova.compute.manager [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 644.496551] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6571e993-84bc-42f5-963a-385b6c0273c6 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "7c48e5b6-eb5f-4c35-9966-1cfc69182595" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.042s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.655155] env[61970]: DEBUG nova.scheduler.client.report [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 644.920036] env[61970]: DEBUG nova.compute.manager [req-1398a0c3-f53a-4c41-8877-88ed49a206bb req-9ce6592a-ede8-4b3e-a462-7fc79a4474b5 service nova] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Received event network-changed-58742262-bf56-4023-8e83-5d4ddd05cfcb {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 644.920343] env[61970]: DEBUG nova.compute.manager [req-1398a0c3-f53a-4c41-8877-88ed49a206bb req-9ce6592a-ede8-4b3e-a462-7fc79a4474b5 service nova] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Refreshing instance network info cache due to event network-changed-58742262-bf56-4023-8e83-5d4ddd05cfcb. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 644.920468] env[61970]: DEBUG oslo_concurrency.lockutils [req-1398a0c3-f53a-4c41-8877-88ed49a206bb req-9ce6592a-ede8-4b3e-a462-7fc79a4474b5 service nova] Acquiring lock "refresh_cache-de9fde95-3ad2-414a-b561-fe4ccc1dc9ba" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.920665] env[61970]: DEBUG oslo_concurrency.lockutils [req-1398a0c3-f53a-4c41-8877-88ed49a206bb req-9ce6592a-ede8-4b3e-a462-7fc79a4474b5 service nova] Acquired lock "refresh_cache-de9fde95-3ad2-414a-b561-fe4ccc1dc9ba" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.920759] env[61970]: DEBUG nova.network.neutron [req-1398a0c3-f53a-4c41-8877-88ed49a206bb req-9ce6592a-ede8-4b3e-a462-7fc79a4474b5 service nova] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Refreshing network info cache for port 58742262-bf56-4023-8e83-5d4ddd05cfcb {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 644.999828] env[61970]: DEBUG nova.compute.manager [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 645.060679] env[61970]: ERROR nova.compute.manager [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 58742262-bf56-4023-8e83-5d4ddd05cfcb, please check neutron logs for more information. [ 645.060679] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 645.060679] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.060679] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 645.060679] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.060679] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 645.060679] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.060679] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 645.060679] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.060679] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 645.060679] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.060679] env[61970]: ERROR nova.compute.manager raise self.value [ 645.060679] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.060679] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 645.060679] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.060679] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 645.061103] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.061103] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 645.061103] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 58742262-bf56-4023-8e83-5d4ddd05cfcb, please check neutron logs for more information. [ 645.061103] env[61970]: ERROR nova.compute.manager [ 645.061103] env[61970]: Traceback (most recent call last): [ 645.061103] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 645.061103] env[61970]: listener.cb(fileno) [ 645.061103] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.061103] env[61970]: result = function(*args, **kwargs) [ 645.061103] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 645.061103] env[61970]: return func(*args, **kwargs) [ 645.061103] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.061103] env[61970]: raise e [ 645.061103] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.061103] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 645.061103] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.061103] env[61970]: created_port_ids = self._update_ports_for_instance( [ 645.061103] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.061103] env[61970]: with excutils.save_and_reraise_exception(): [ 645.061103] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.061103] env[61970]: self.force_reraise() [ 645.061103] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.061103] env[61970]: raise self.value [ 645.061103] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.061103] env[61970]: updated_port = self._update_port( [ 645.061103] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.061103] env[61970]: _ensure_no_port_binding_failure(port) [ 645.061103] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.061103] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 645.061755] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 58742262-bf56-4023-8e83-5d4ddd05cfcb, please check neutron logs for more information. [ 645.061755] env[61970]: Removing descriptor: 17 [ 645.163874] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.164524] env[61970]: ERROR nova.compute.manager [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f3160862-d364-49dd-8639-282b45d418b5, please check neutron logs for more information. [ 645.164524] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Traceback (most recent call last): [ 645.164524] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 645.164524] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] self.driver.spawn(context, instance, image_meta, [ 645.164524] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 645.164524] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 645.164524] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 645.164524] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] vm_ref = self.build_virtual_machine(instance, [ 645.164524] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 645.164524] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] vif_infos = vmwarevif.get_vif_info(self._session, [ 645.164524] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 645.164864] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] for vif in network_info: [ 645.164864] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 645.164864] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] return self._sync_wrapper(fn, *args, **kwargs) [ 645.164864] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 645.164864] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] self.wait() [ 645.164864] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 645.164864] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] self[:] = self._gt.wait() [ 645.164864] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 645.164864] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] return self._exit_event.wait() [ 645.164864] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 645.164864] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] result = hub.switch() [ 645.164864] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 645.164864] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] return self.greenlet.switch() [ 645.165241] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.165241] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] result = function(*args, **kwargs) [ 645.165241] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 645.165241] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] return func(*args, **kwargs) [ 645.165241] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.165241] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] raise e [ 645.165241] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.165241] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] nwinfo = self.network_api.allocate_for_instance( [ 645.165241] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.165241] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] created_port_ids = self._update_ports_for_instance( [ 645.165241] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.165241] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] with excutils.save_and_reraise_exception(): [ 645.165241] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.165610] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] self.force_reraise() [ 645.165610] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.165610] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] raise self.value [ 645.165610] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.165610] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] updated_port = self._update_port( [ 645.165610] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.165610] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] _ensure_no_port_binding_failure(port) [ 645.165610] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.165610] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] raise exception.PortBindingFailed(port_id=port['id']) [ 645.165610] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] nova.exception.PortBindingFailed: Binding failed for port f3160862-d364-49dd-8639-282b45d418b5, please check neutron logs for more information. [ 645.165610] env[61970]: ERROR nova.compute.manager [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] [ 645.165936] env[61970]: DEBUG nova.compute.utils [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Binding failed for port f3160862-d364-49dd-8639-282b45d418b5, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 645.166472] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.423s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.168034] env[61970]: INFO nova.compute.claims [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 645.171500] env[61970]: DEBUG nova.compute.manager [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Build of instance a6a3b8fe-6911-4d8f-bd3e-8044c47a541b was re-scheduled: Binding failed for port f3160862-d364-49dd-8639-282b45d418b5, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 645.171894] env[61970]: DEBUG nova.compute.manager [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 645.173028] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Acquiring lock "refresh_cache-a6a3b8fe-6911-4d8f-bd3e-8044c47a541b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.173028] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Acquired lock "refresh_cache-a6a3b8fe-6911-4d8f-bd3e-8044c47a541b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.173028] env[61970]: DEBUG nova.network.neutron [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 645.188819] env[61970]: DEBUG nova.compute.manager [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 645.218257] env[61970]: DEBUG nova.virt.hardware [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:38:50Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='c300c635-c3de-425c-93b4-132ae0a55594',id=27,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1265825726',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 645.218498] env[61970]: DEBUG nova.virt.hardware [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 645.218667] env[61970]: DEBUG nova.virt.hardware [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 645.218870] env[61970]: DEBUG nova.virt.hardware [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 645.218997] env[61970]: DEBUG nova.virt.hardware [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 645.219714] env[61970]: DEBUG nova.virt.hardware [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 645.219952] env[61970]: DEBUG nova.virt.hardware [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 645.220252] env[61970]: DEBUG nova.virt.hardware [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 645.220449] env[61970]: DEBUG nova.virt.hardware [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 645.220619] env[61970]: DEBUG nova.virt.hardware [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 645.220794] env[61970]: DEBUG nova.virt.hardware [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 645.221913] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eb41967-736b-4a63-809c-88314e3f294b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.230960] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4058566b-74d9-4410-b886-26f803d293b5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.252074] env[61970]: ERROR nova.compute.manager [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 58742262-bf56-4023-8e83-5d4ddd05cfcb, please check neutron logs for more information. [ 645.252074] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Traceback (most recent call last): [ 645.252074] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 645.252074] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] yield resources [ 645.252074] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 645.252074] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] self.driver.spawn(context, instance, image_meta, [ 645.252074] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 645.252074] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 645.252074] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 645.252074] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] vm_ref = self.build_virtual_machine(instance, [ 645.252074] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 645.252498] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 645.252498] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 645.252498] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] for vif in network_info: [ 645.252498] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 645.252498] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] return self._sync_wrapper(fn, *args, **kwargs) [ 645.252498] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 645.252498] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] self.wait() [ 645.252498] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 645.252498] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] self[:] = self._gt.wait() [ 645.252498] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 645.252498] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] return self._exit_event.wait() [ 645.252498] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 645.252498] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] current.throw(*self._exc) [ 645.252860] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.252860] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] result = function(*args, **kwargs) [ 645.252860] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 645.252860] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] return func(*args, **kwargs) [ 645.252860] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.252860] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] raise e [ 645.252860] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.252860] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] nwinfo = self.network_api.allocate_for_instance( [ 645.252860] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.252860] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] created_port_ids = self._update_ports_for_instance( [ 645.252860] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.252860] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] with excutils.save_and_reraise_exception(): [ 645.252860] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.253240] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] self.force_reraise() [ 645.253240] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.253240] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] raise self.value [ 645.253240] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.253240] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] updated_port = self._update_port( [ 645.253240] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.253240] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] _ensure_no_port_binding_failure(port) [ 645.253240] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.253240] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] raise exception.PortBindingFailed(port_id=port['id']) [ 645.253240] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] nova.exception.PortBindingFailed: Binding failed for port 58742262-bf56-4023-8e83-5d4ddd05cfcb, please check neutron logs for more information. [ 645.253240] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] [ 645.253240] env[61970]: INFO nova.compute.manager [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Terminating instance [ 645.253577] env[61970]: DEBUG oslo_concurrency.lockutils [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Acquiring lock "refresh_cache-de9fde95-3ad2-414a-b561-fe4ccc1dc9ba" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.442793] env[61970]: DEBUG nova.network.neutron [req-1398a0c3-f53a-4c41-8877-88ed49a206bb req-9ce6592a-ede8-4b3e-a462-7fc79a4474b5 service nova] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.499124] env[61970]: DEBUG nova.network.neutron [req-1398a0c3-f53a-4c41-8877-88ed49a206bb req-9ce6592a-ede8-4b3e-a462-7fc79a4474b5 service nova] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.517807] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.688281] env[61970]: DEBUG nova.network.neutron [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.764842] env[61970]: DEBUG nova.network.neutron [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.810931] env[61970]: DEBUG oslo_concurrency.lockutils [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "fd4274ed-bf9a-45ab-a350-a8ebbc5ad404" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.811130] env[61970]: DEBUG oslo_concurrency.lockutils [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "fd4274ed-bf9a-45ab-a350-a8ebbc5ad404" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.002260] env[61970]: DEBUG oslo_concurrency.lockutils [req-1398a0c3-f53a-4c41-8877-88ed49a206bb req-9ce6592a-ede8-4b3e-a462-7fc79a4474b5 service nova] Releasing lock "refresh_cache-de9fde95-3ad2-414a-b561-fe4ccc1dc9ba" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.002756] env[61970]: DEBUG oslo_concurrency.lockutils [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Acquired lock "refresh_cache-de9fde95-3ad2-414a-b561-fe4ccc1dc9ba" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.003438] env[61970]: DEBUG nova.network.neutron [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 646.269274] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Releasing lock "refresh_cache-a6a3b8fe-6911-4d8f-bd3e-8044c47a541b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.269495] env[61970]: DEBUG nova.compute.manager [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 646.269676] env[61970]: DEBUG nova.compute.manager [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 646.269838] env[61970]: DEBUG nova.network.neutron [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 646.285399] env[61970]: DEBUG nova.network.neutron [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.482147] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2767864e-0cf8-4610-82bb-20a38cdf77b9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.488435] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18cab697-d8a5-491c-b3c3-0cdd5dec91f7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.521316] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be2f17cf-93f7-4808-b379-8b92e4202338 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.529179] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b233a71c-dcdc-4fa7-8a21-6d5aa28bcb26 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.534997] env[61970]: DEBUG nova.network.neutron [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.544571] env[61970]: DEBUG nova.compute.provider_tree [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 646.589473] env[61970]: DEBUG nova.network.neutron [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.788021] env[61970]: DEBUG nova.network.neutron [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.969355] env[61970]: DEBUG nova.compute.manager [req-838216a0-b82c-46f7-bab8-480e6a2a8563 req-f2195105-569b-4fd9-a591-50196d5312d6 service nova] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Received event network-vif-deleted-58742262-bf56-4023-8e83-5d4ddd05cfcb {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 647.049795] env[61970]: DEBUG nova.scheduler.client.report [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 647.092475] env[61970]: DEBUG oslo_concurrency.lockutils [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Releasing lock "refresh_cache-de9fde95-3ad2-414a-b561-fe4ccc1dc9ba" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.092876] env[61970]: DEBUG nova.compute.manager [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 647.093412] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 647.093412] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-50c78f4e-8e80-4f92-8368-7b39569f83d9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.102253] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af1e0075-79d1-4107-b7e7-ef6642186e29 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.122867] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance de9fde95-3ad2-414a-b561-fe4ccc1dc9ba could not be found. [ 647.123083] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 647.123262] env[61970]: INFO nova.compute.manager [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Took 0.03 seconds to destroy the instance on the hypervisor. [ 647.123498] env[61970]: DEBUG oslo.service.loopingcall [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 647.123701] env[61970]: DEBUG nova.compute.manager [-] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 647.123841] env[61970]: DEBUG nova.network.neutron [-] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 647.143317] env[61970]: DEBUG nova.network.neutron [-] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 647.290444] env[61970]: INFO nova.compute.manager [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] [instance: a6a3b8fe-6911-4d8f-bd3e-8044c47a541b] Took 1.02 seconds to deallocate network for instance. [ 647.554870] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.555371] env[61970]: DEBUG nova.compute.manager [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 647.558176] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.829s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.558352] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.558503] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61970) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 647.558831] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.161s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.563385] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a67031c-0ce8-45d2-a3ee-afd6f82b40ec {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.572306] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3534062-0f83-477a-a36c-5a5b1b1d6043 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.587944] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ad9ad5-278a-47a7-909e-33ff7f2ab545 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.596024] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53dbeca4-d8e7-41a1-8aa8-1a857c8c6dea {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.624535] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181458MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61970) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 647.624657] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.645395] env[61970]: DEBUG nova.network.neutron [-] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.068107] env[61970]: DEBUG nova.compute.utils [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 648.070657] env[61970]: DEBUG nova.compute.manager [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 648.070925] env[61970]: DEBUG nova.network.neutron [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 648.110139] env[61970]: DEBUG nova.policy [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'daa274597c93401594315dc11492c250', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ca3a5b33410c4428a14516d3d63dcc26', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 648.148034] env[61970]: INFO nova.compute.manager [-] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Took 1.02 seconds to deallocate network for instance. [ 648.152310] env[61970]: DEBUG nova.compute.claims [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 648.152494] env[61970]: DEBUG oslo_concurrency.lockutils [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.320449] env[61970]: INFO nova.scheduler.client.report [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Deleted allocations for instance a6a3b8fe-6911-4d8f-bd3e-8044c47a541b [ 648.424806] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb8868e-1ee9-4ca0-8dd4-dcb722e26f81 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.435870] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b0b4c9-fef3-4327-b087-1fd1e276ab35 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.470825] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-588629a3-4c20-4445-b256-7536578afac7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.478916] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7fdf63f-5ef3-436e-a9c8-a5a18ce02197 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.485214] env[61970]: DEBUG nova.network.neutron [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Successfully created port: 4d3960a5-7523-42c8-9e11-6bfa1e9d8a51 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 648.494963] env[61970]: DEBUG nova.compute.provider_tree [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 648.574245] env[61970]: DEBUG nova.compute.manager [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 648.830121] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b374446a-22c7-4a0b-8035-f9d1a63b602f tempest-ImagesOneServerTestJSON-1341450677 tempest-ImagesOneServerTestJSON-1341450677-project-member] Lock "a6a3b8fe-6911-4d8f-bd3e-8044c47a541b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.057s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.998505] env[61970]: DEBUG nova.scheduler.client.report [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.080913] env[61970]: INFO nova.virt.block_device [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Booting with volume 5872d0cc-3d4c-4007-9609-6d6dc3474779 at /dev/sda [ 649.140859] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a35bf7eb-7542-4410-9fd9-2a0a291d8467 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.148215] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef7129ec-af87-4f87-bac3-03474b4de749 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.170949] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c5e91bd8-d519-48b9-81a2-87975b0ab0dd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.178977] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70906a91-2edd-4c3f-baff-2b89a1df02a4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.201800] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2ee95c2-b20f-4a10-94b6-e3290bc27e8d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.208554] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8d2c0b-d699-46b5-8663-5a86932bbe06 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.222562] env[61970]: DEBUG nova.virt.block_device [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Updating existing volume attachment record: a06732f2-789f-4d61-8b69-ae1e307509bd {{(pid=61970) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 649.269589] env[61970]: DEBUG nova.compute.manager [req-4928b41e-8859-47c1-9365-c01acaed583f req-8bb33ed9-0dda-44fa-a244-86ed209cdced service nova] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Received event network-changed-4d3960a5-7523-42c8-9e11-6bfa1e9d8a51 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 649.269795] env[61970]: DEBUG nova.compute.manager [req-4928b41e-8859-47c1-9365-c01acaed583f req-8bb33ed9-0dda-44fa-a244-86ed209cdced service nova] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Refreshing instance network info cache due to event network-changed-4d3960a5-7523-42c8-9e11-6bfa1e9d8a51. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 649.270099] env[61970]: DEBUG oslo_concurrency.lockutils [req-4928b41e-8859-47c1-9365-c01acaed583f req-8bb33ed9-0dda-44fa-a244-86ed209cdced service nova] Acquiring lock "refresh_cache-293b6863-510c-493e-8c9a-af53ad80fc89" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.272034] env[61970]: DEBUG oslo_concurrency.lockutils [req-4928b41e-8859-47c1-9365-c01acaed583f req-8bb33ed9-0dda-44fa-a244-86ed209cdced service nova] Acquired lock "refresh_cache-293b6863-510c-493e-8c9a-af53ad80fc89" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.272034] env[61970]: DEBUG nova.network.neutron [req-4928b41e-8859-47c1-9365-c01acaed583f req-8bb33ed9-0dda-44fa-a244-86ed209cdced service nova] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Refreshing network info cache for port 4d3960a5-7523-42c8-9e11-6bfa1e9d8a51 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 649.333279] env[61970]: DEBUG nova.compute.manager [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 649.504772] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.945s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.505527] env[61970]: ERROR nova.compute.manager [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f82e6809-c983-45f2-b9bf-02b19bfb6043, please check neutron logs for more information. [ 649.505527] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Traceback (most recent call last): [ 649.505527] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 649.505527] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] self.driver.spawn(context, instance, image_meta, [ 649.505527] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 649.505527] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 649.505527] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 649.505527] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] vm_ref = self.build_virtual_machine(instance, [ 649.505527] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 649.505527] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 649.505527] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 649.506085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] for vif in network_info: [ 649.506085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 649.506085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] return self._sync_wrapper(fn, *args, **kwargs) [ 649.506085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 649.506085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] self.wait() [ 649.506085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 649.506085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] self[:] = self._gt.wait() [ 649.506085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 649.506085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] return self._exit_event.wait() [ 649.506085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 649.506085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] result = hub.switch() [ 649.506085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 649.506085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] return self.greenlet.switch() [ 649.508085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.508085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] result = function(*args, **kwargs) [ 649.508085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 649.508085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] return func(*args, **kwargs) [ 649.508085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 649.508085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] raise e [ 649.508085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.508085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] nwinfo = self.network_api.allocate_for_instance( [ 649.508085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.508085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] created_port_ids = self._update_ports_for_instance( [ 649.508085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.508085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] with excutils.save_and_reraise_exception(): [ 649.508085] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.508521] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] self.force_reraise() [ 649.508521] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.508521] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] raise self.value [ 649.508521] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.508521] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] updated_port = self._update_port( [ 649.508521] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.508521] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] _ensure_no_port_binding_failure(port) [ 649.508521] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.508521] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] raise exception.PortBindingFailed(port_id=port['id']) [ 649.508521] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] nova.exception.PortBindingFailed: Binding failed for port f82e6809-c983-45f2-b9bf-02b19bfb6043, please check neutron logs for more information. [ 649.508521] env[61970]: ERROR nova.compute.manager [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] [ 649.508804] env[61970]: DEBUG nova.compute.utils [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Binding failed for port f82e6809-c983-45f2-b9bf-02b19bfb6043, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 649.508804] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.769s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.512269] env[61970]: DEBUG nova.compute.manager [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Build of instance 65dba389-321c-478d-9555-bcd9c318f6d0 was re-scheduled: Binding failed for port f82e6809-c983-45f2-b9bf-02b19bfb6043, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 649.512744] env[61970]: DEBUG nova.compute.manager [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 649.512977] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Acquiring lock "refresh_cache-65dba389-321c-478d-9555-bcd9c318f6d0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.513141] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Acquired lock "refresh_cache-65dba389-321c-478d-9555-bcd9c318f6d0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.513377] env[61970]: DEBUG nova.network.neutron [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 649.710064] env[61970]: ERROR nova.compute.manager [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4d3960a5-7523-42c8-9e11-6bfa1e9d8a51, please check neutron logs for more information. [ 649.710064] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 649.710064] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.710064] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 649.710064] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.710064] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 649.710064] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.710064] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 649.710064] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.710064] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 649.710064] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.710064] env[61970]: ERROR nova.compute.manager raise self.value [ 649.710064] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.710064] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 649.710064] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.710064] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 649.710629] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.710629] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 649.710629] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4d3960a5-7523-42c8-9e11-6bfa1e9d8a51, please check neutron logs for more information. [ 649.710629] env[61970]: ERROR nova.compute.manager [ 649.710629] env[61970]: Traceback (most recent call last): [ 649.710629] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 649.710629] env[61970]: listener.cb(fileno) [ 649.710629] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.710629] env[61970]: result = function(*args, **kwargs) [ 649.710629] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 649.710629] env[61970]: return func(*args, **kwargs) [ 649.710629] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 649.710629] env[61970]: raise e [ 649.710629] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.710629] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 649.710629] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.710629] env[61970]: created_port_ids = self._update_ports_for_instance( [ 649.710629] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.710629] env[61970]: with excutils.save_and_reraise_exception(): [ 649.710629] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.710629] env[61970]: self.force_reraise() [ 649.710629] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.710629] env[61970]: raise self.value [ 649.710629] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.710629] env[61970]: updated_port = self._update_port( [ 649.710629] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.710629] env[61970]: _ensure_no_port_binding_failure(port) [ 649.710629] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.710629] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 649.711639] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 4d3960a5-7523-42c8-9e11-6bfa1e9d8a51, please check neutron logs for more information. [ 649.711639] env[61970]: Removing descriptor: 17 [ 649.789092] env[61970]: DEBUG nova.network.neutron [req-4928b41e-8859-47c1-9365-c01acaed583f req-8bb33ed9-0dda-44fa-a244-86ed209cdced service nova] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.860138] env[61970]: DEBUG oslo_concurrency.lockutils [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.932473] env[61970]: DEBUG nova.network.neutron [req-4928b41e-8859-47c1-9365-c01acaed583f req-8bb33ed9-0dda-44fa-a244-86ed209cdced service nova] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.054536] env[61970]: DEBUG nova.network.neutron [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.374293] env[61970]: DEBUG nova.network.neutron [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.403971] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12746301-5bb6-4573-afeb-9f843dde8f75 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.411400] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d32fecb1-75f5-4116-b2b2-d06a88d8ee93 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.441850] env[61970]: DEBUG oslo_concurrency.lockutils [req-4928b41e-8859-47c1-9365-c01acaed583f req-8bb33ed9-0dda-44fa-a244-86ed209cdced service nova] Releasing lock "refresh_cache-293b6863-510c-493e-8c9a-af53ad80fc89" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.442879] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-452c4505-70bd-47c5-803b-d060a6b12e33 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.452393] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6122c9fd-9a34-4648-a508-71e172f72dfd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.469685] env[61970]: DEBUG nova.compute.provider_tree [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.877134] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Releasing lock "refresh_cache-65dba389-321c-478d-9555-bcd9c318f6d0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.877399] env[61970]: DEBUG nova.compute.manager [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 650.877608] env[61970]: DEBUG nova.compute.manager [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 650.877782] env[61970]: DEBUG nova.network.neutron [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 650.896066] env[61970]: DEBUG nova.network.neutron [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.972754] env[61970]: DEBUG nova.scheduler.client.report [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 651.298412] env[61970]: DEBUG nova.compute.manager [req-49ea7edb-8f18-4a65-aee8-f67dc3eaef2e req-43f62a5b-4e15-4c07-b2ca-36ed1c3b7dcf service nova] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Received event network-vif-deleted-4d3960a5-7523-42c8-9e11-6bfa1e9d8a51 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 651.358525] env[61970]: DEBUG nova.compute.manager [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 651.359069] env[61970]: DEBUG nova.virt.hardware [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 651.359274] env[61970]: DEBUG nova.virt.hardware [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 651.359422] env[61970]: DEBUG nova.virt.hardware [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 651.359592] env[61970]: DEBUG nova.virt.hardware [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 651.359739] env[61970]: DEBUG nova.virt.hardware [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 651.359875] env[61970]: DEBUG nova.virt.hardware [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 651.360081] env[61970]: DEBUG nova.virt.hardware [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 651.360239] env[61970]: DEBUG nova.virt.hardware [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 651.360395] env[61970]: DEBUG nova.virt.hardware [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 651.360549] env[61970]: DEBUG nova.virt.hardware [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 651.360712] env[61970]: DEBUG nova.virt.hardware [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 651.363247] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a9ad52-21d6-4ac7-86ee-8cefe014cc42 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.370667] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61733e4a-3cdd-4ac4-b274-ef99f900c888 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.387563] env[61970]: ERROR nova.compute.manager [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4d3960a5-7523-42c8-9e11-6bfa1e9d8a51, please check neutron logs for more information. [ 651.387563] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Traceback (most recent call last): [ 651.387563] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 651.387563] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] yield resources [ 651.387563] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 651.387563] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] self.driver.spawn(context, instance, image_meta, [ 651.387563] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 651.387563] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] self._vmops.spawn(context, instance, image_meta, injected_files, [ 651.387563] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 651.387563] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] vm_ref = self.build_virtual_machine(instance, [ 651.387563] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 651.388253] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] vif_infos = vmwarevif.get_vif_info(self._session, [ 651.388253] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 651.388253] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] for vif in network_info: [ 651.388253] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 651.388253] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] return self._sync_wrapper(fn, *args, **kwargs) [ 651.388253] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 651.388253] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] self.wait() [ 651.388253] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 651.388253] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] self[:] = self._gt.wait() [ 651.388253] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 651.388253] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] return self._exit_event.wait() [ 651.388253] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 651.388253] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] current.throw(*self._exc) [ 651.388647] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.388647] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] result = function(*args, **kwargs) [ 651.388647] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 651.388647] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] return func(*args, **kwargs) [ 651.388647] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.388647] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] raise e [ 651.388647] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.388647] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] nwinfo = self.network_api.allocate_for_instance( [ 651.388647] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.388647] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] created_port_ids = self._update_ports_for_instance( [ 651.388647] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.388647] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] with excutils.save_and_reraise_exception(): [ 651.388647] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.389034] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] self.force_reraise() [ 651.389034] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.389034] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] raise self.value [ 651.389034] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.389034] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] updated_port = self._update_port( [ 651.389034] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.389034] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] _ensure_no_port_binding_failure(port) [ 651.389034] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.389034] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] raise exception.PortBindingFailed(port_id=port['id']) [ 651.389034] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] nova.exception.PortBindingFailed: Binding failed for port 4d3960a5-7523-42c8-9e11-6bfa1e9d8a51, please check neutron logs for more information. [ 651.389034] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] [ 651.389034] env[61970]: INFO nova.compute.manager [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Terminating instance [ 651.390398] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Acquiring lock "refresh_cache-293b6863-510c-493e-8c9a-af53ad80fc89" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.390487] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Acquired lock "refresh_cache-293b6863-510c-493e-8c9a-af53ad80fc89" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.390812] env[61970]: DEBUG nova.network.neutron [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 651.399962] env[61970]: DEBUG nova.network.neutron [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.478585] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.969s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.479477] env[61970]: ERROR nova.compute.manager [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 73c584a8-7f1c-4abd-8b28-60519162a678, please check neutron logs for more information. [ 651.479477] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Traceback (most recent call last): [ 651.479477] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 651.479477] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] self.driver.spawn(context, instance, image_meta, [ 651.479477] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 651.479477] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 651.479477] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 651.479477] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] vm_ref = self.build_virtual_machine(instance, [ 651.479477] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 651.479477] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] vif_infos = vmwarevif.get_vif_info(self._session, [ 651.479477] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 651.479813] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] for vif in network_info: [ 651.479813] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 651.479813] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] return self._sync_wrapper(fn, *args, **kwargs) [ 651.479813] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 651.479813] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] self.wait() [ 651.479813] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 651.479813] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] self[:] = self._gt.wait() [ 651.479813] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 651.479813] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] return self._exit_event.wait() [ 651.479813] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 651.479813] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] result = hub.switch() [ 651.479813] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 651.479813] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] return self.greenlet.switch() [ 651.480152] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.480152] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] result = function(*args, **kwargs) [ 651.480152] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 651.480152] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] return func(*args, **kwargs) [ 651.480152] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.480152] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] raise e [ 651.480152] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.480152] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] nwinfo = self.network_api.allocate_for_instance( [ 651.480152] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.480152] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] created_port_ids = self._update_ports_for_instance( [ 651.480152] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.480152] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] with excutils.save_and_reraise_exception(): [ 651.480152] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.480449] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] self.force_reraise() [ 651.480449] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.480449] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] raise self.value [ 651.480449] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.480449] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] updated_port = self._update_port( [ 651.480449] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.480449] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] _ensure_no_port_binding_failure(port) [ 651.480449] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.480449] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] raise exception.PortBindingFailed(port_id=port['id']) [ 651.480449] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] nova.exception.PortBindingFailed: Binding failed for port 73c584a8-7f1c-4abd-8b28-60519162a678, please check neutron logs for more information. [ 651.480449] env[61970]: ERROR nova.compute.manager [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] [ 651.480719] env[61970]: DEBUG nova.compute.utils [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Binding failed for port 73c584a8-7f1c-4abd-8b28-60519162a678, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 651.480879] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.521s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.482333] env[61970]: INFO nova.compute.claims [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 651.486058] env[61970]: DEBUG nova.compute.manager [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Build of instance 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4 was re-scheduled: Binding failed for port 73c584a8-7f1c-4abd-8b28-60519162a678, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 651.486516] env[61970]: DEBUG nova.compute.manager [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 651.486746] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Acquiring lock "refresh_cache-0c45c5c0-f8db-4631-ba9a-271d8f9ddff4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.486894] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Acquired lock "refresh_cache-0c45c5c0-f8db-4631-ba9a-271d8f9ddff4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.487095] env[61970]: DEBUG nova.network.neutron [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 651.903539] env[61970]: INFO nova.compute.manager [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 65dba389-321c-478d-9555-bcd9c318f6d0] Took 1.03 seconds to deallocate network for instance. [ 651.918786] env[61970]: DEBUG nova.network.neutron [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.016825] env[61970]: DEBUG nova.network.neutron [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.095606] env[61970]: DEBUG nova.network.neutron [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.150944] env[61970]: DEBUG nova.network.neutron [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.597797] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Releasing lock "refresh_cache-293b6863-510c-493e-8c9a-af53ad80fc89" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.598365] env[61970]: DEBUG nova.compute.manager [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 652.598729] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d6250820-4ab0-48a3-ad1d-5d296acd9447 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.607790] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2607f45-2cc3-46d6-9f18-71d83daea41b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.632386] env[61970]: WARNING nova.virt.vmwareapi.driver [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 293b6863-510c-493e-8c9a-af53ad80fc89 could not be found. [ 652.632614] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 652.635121] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d846ea82-db97-4aad-8227-545f4a650af1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.643933] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2be277-6feb-4ca8-b180-85c20611bfd3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.656680] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Releasing lock "refresh_cache-0c45c5c0-f8db-4631-ba9a-271d8f9ddff4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.656976] env[61970]: DEBUG nova.compute.manager [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 652.657183] env[61970]: DEBUG nova.compute.manager [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 652.657350] env[61970]: DEBUG nova.network.neutron [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 652.671555] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 293b6863-510c-493e-8c9a-af53ad80fc89 could not be found. [ 652.671844] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 652.672050] env[61970]: INFO nova.compute.manager [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Took 0.07 seconds to destroy the instance on the hypervisor. [ 652.672410] env[61970]: DEBUG oslo.service.loopingcall [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 652.677410] env[61970]: DEBUG nova.compute.manager [-] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 652.677410] env[61970]: DEBUG nova.network.neutron [-] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 652.678393] env[61970]: DEBUG nova.network.neutron [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.689026] env[61970]: DEBUG nova.network.neutron [-] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.891021] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdb43c02-8995-4a9f-ad2d-d4c933b2b84f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.899704] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f89ba10-1168-4e5e-bfd0-4bc96aa36726 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.933372] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c5de00-04ad-4442-ab05-5e525418306c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.940728] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e6cecb-d224-44f1-9658-014dde71737c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.953934] env[61970]: DEBUG nova.compute.provider_tree [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.955948] env[61970]: INFO nova.scheduler.client.report [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Deleted allocations for instance 65dba389-321c-478d-9555-bcd9c318f6d0 [ 653.180462] env[61970]: DEBUG nova.network.neutron [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.191637] env[61970]: DEBUG nova.network.neutron [-] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.469485] env[61970]: DEBUG nova.scheduler.client.report [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 653.473366] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c678d370-3b20-4622-b670-2fb564ddcda6 tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Lock "65dba389-321c-478d-9555-bcd9c318f6d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 87.684s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.685618] env[61970]: INFO nova.compute.manager [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] [instance: 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4] Took 1.03 seconds to deallocate network for instance. [ 653.695516] env[61970]: INFO nova.compute.manager [-] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Took 1.02 seconds to deallocate network for instance. [ 653.974953] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.494s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.977198] env[61970]: DEBUG nova.compute.manager [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 653.979766] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.261s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.983754] env[61970]: DEBUG nova.compute.manager [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 654.261691] env[61970]: INFO nova.compute.manager [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Took 0.57 seconds to detach 1 volumes for instance. [ 654.263985] env[61970]: DEBUG nova.compute.claims [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 654.264548] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.490867] env[61970]: DEBUG nova.compute.utils [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 654.497745] env[61970]: DEBUG nova.compute.manager [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 654.497929] env[61970]: DEBUG nova.network.neutron [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 654.521269] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.556135] env[61970]: DEBUG nova.policy [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '74c51f80b8bd42d1a1d9c51e3d578b95', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '41255ef8a2f247ccada14156b1b9224a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 654.728048] env[61970]: INFO nova.scheduler.client.report [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Deleted allocations for instance 0c45c5c0-f8db-4631-ba9a-271d8f9ddff4 [ 654.903029] env[61970]: DEBUG nova.network.neutron [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Successfully created port: 6d06502e-9245-444a-8e44-afa3e0c2874b {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 654.947559] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a0ffb0-238e-410a-ac94-9140686ec53f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.957558] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-514a862b-3926-4346-b2cf-bcced341fc11 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.998100] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b5cbda-8884-4bdd-8b0d-eb53cc0ecc42 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.005127] env[61970]: DEBUG nova.compute.manager [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 655.008716] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97679a69-36c9-4e8c-bfff-ae137a7a7776 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.025223] env[61970]: DEBUG nova.compute.provider_tree [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.240506] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ae9356bf-7e6b-4673-9292-9fb69620356e tempest-ServersTestFqdnHostnames-1782808197 tempest-ServersTestFqdnHostnames-1782808197-project-member] Lock "0c45c5c0-f8db-4631-ba9a-271d8f9ddff4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.943s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.410610] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "e74c9188-4f8a-4071-bc4f-c2be91fec3f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.411052] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "e74c9188-4f8a-4071-bc4f-c2be91fec3f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.435827] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "01ee68e2-4310-477e-8adb-6ad67d6a65e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.436056] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "01ee68e2-4310-477e-8adb-6ad67d6a65e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.529745] env[61970]: DEBUG nova.scheduler.client.report [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.743078] env[61970]: DEBUG nova.compute.manager [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 655.943243] env[61970]: DEBUG nova.compute.manager [req-f219c466-3fd8-4418-a5bd-50ca05fe9df6 req-81ebd564-55b1-46c2-9df5-063f097ee10f service nova] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Received event network-changed-6d06502e-9245-444a-8e44-afa3e0c2874b {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 655.943446] env[61970]: DEBUG nova.compute.manager [req-f219c466-3fd8-4418-a5bd-50ca05fe9df6 req-81ebd564-55b1-46c2-9df5-063f097ee10f service nova] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Refreshing instance network info cache due to event network-changed-6d06502e-9245-444a-8e44-afa3e0c2874b. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 655.943617] env[61970]: DEBUG oslo_concurrency.lockutils [req-f219c466-3fd8-4418-a5bd-50ca05fe9df6 req-81ebd564-55b1-46c2-9df5-063f097ee10f service nova] Acquiring lock "refresh_cache-77460934-215e-4294-8eb3-a380dfbacddd" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.943985] env[61970]: DEBUG oslo_concurrency.lockutils [req-f219c466-3fd8-4418-a5bd-50ca05fe9df6 req-81ebd564-55b1-46c2-9df5-063f097ee10f service nova] Acquired lock "refresh_cache-77460934-215e-4294-8eb3-a380dfbacddd" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.944174] env[61970]: DEBUG nova.network.neutron [req-f219c466-3fd8-4418-a5bd-50ca05fe9df6 req-81ebd564-55b1-46c2-9df5-063f097ee10f service nova] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Refreshing network info cache for port 6d06502e-9245-444a-8e44-afa3e0c2874b {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 656.019575] env[61970]: DEBUG nova.compute.manager [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 656.036392] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.056s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.037073] env[61970]: ERROR nova.compute.manager [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a664bf5a-c723-43e0-a1c2-f19168e6054e, please check neutron logs for more information. [ 656.037073] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Traceback (most recent call last): [ 656.037073] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.037073] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] self.driver.spawn(context, instance, image_meta, [ 656.037073] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 656.037073] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.037073] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.037073] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] vm_ref = self.build_virtual_machine(instance, [ 656.037073] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.037073] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.037073] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.037433] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] for vif in network_info: [ 656.037433] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 656.037433] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] return self._sync_wrapper(fn, *args, **kwargs) [ 656.037433] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 656.037433] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] self.wait() [ 656.037433] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 656.037433] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] self[:] = self._gt.wait() [ 656.037433] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.037433] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] return self._exit_event.wait() [ 656.037433] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 656.037433] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] result = hub.switch() [ 656.037433] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 656.037433] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] return self.greenlet.switch() [ 656.037848] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.037848] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] result = function(*args, **kwargs) [ 656.037848] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 656.037848] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] return func(*args, **kwargs) [ 656.037848] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.037848] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] raise e [ 656.037848] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.037848] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] nwinfo = self.network_api.allocate_for_instance( [ 656.037848] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.037848] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] created_port_ids = self._update_ports_for_instance( [ 656.037848] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.037848] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] with excutils.save_and_reraise_exception(): [ 656.037848] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.038220] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] self.force_reraise() [ 656.038220] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.038220] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] raise self.value [ 656.038220] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.038220] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] updated_port = self._update_port( [ 656.038220] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.038220] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] _ensure_no_port_binding_failure(port) [ 656.038220] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.038220] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] raise exception.PortBindingFailed(port_id=port['id']) [ 656.038220] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] nova.exception.PortBindingFailed: Binding failed for port a664bf5a-c723-43e0-a1c2-f19168e6054e, please check neutron logs for more information. [ 656.038220] env[61970]: ERROR nova.compute.manager [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] [ 656.038510] env[61970]: DEBUG nova.compute.utils [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Binding failed for port a664bf5a-c723-43e0-a1c2-f19168e6054e, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 656.044563] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.434s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.047803] env[61970]: DEBUG nova.compute.manager [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Build of instance 816de9f9-a8a9-415a-a6c4-57baea86df51 was re-scheduled: Binding failed for port a664bf5a-c723-43e0-a1c2-f19168e6054e, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 656.048165] env[61970]: DEBUG nova.compute.manager [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 656.050534] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Acquiring lock "refresh_cache-816de9f9-a8a9-415a-a6c4-57baea86df51" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.050534] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Acquired lock "refresh_cache-816de9f9-a8a9-415a-a6c4-57baea86df51" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.050534] env[61970]: DEBUG nova.network.neutron [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 656.058818] env[61970]: DEBUG nova.virt.hardware [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 656.059122] env[61970]: DEBUG nova.virt.hardware [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 656.059346] env[61970]: DEBUG nova.virt.hardware [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 656.059547] env[61970]: DEBUG nova.virt.hardware [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 656.059697] env[61970]: DEBUG nova.virt.hardware [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 656.059849] env[61970]: DEBUG nova.virt.hardware [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 656.060079] env[61970]: DEBUG nova.virt.hardware [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 656.060245] env[61970]: DEBUG nova.virt.hardware [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 656.060416] env[61970]: DEBUG nova.virt.hardware [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 656.060578] env[61970]: DEBUG nova.virt.hardware [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 656.060749] env[61970]: DEBUG nova.virt.hardware [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 656.061753] env[61970]: ERROR nova.compute.manager [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6d06502e-9245-444a-8e44-afa3e0c2874b, please check neutron logs for more information. [ 656.061753] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 656.061753] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.061753] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 656.061753] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.061753] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 656.061753] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.061753] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 656.061753] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.061753] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 656.061753] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.061753] env[61970]: ERROR nova.compute.manager raise self.value [ 656.061753] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.061753] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 656.061753] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.061753] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 656.062232] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.062232] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 656.062232] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6d06502e-9245-444a-8e44-afa3e0c2874b, please check neutron logs for more information. [ 656.062232] env[61970]: ERROR nova.compute.manager [ 656.062232] env[61970]: Traceback (most recent call last): [ 656.062232] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 656.062232] env[61970]: listener.cb(fileno) [ 656.062232] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.062232] env[61970]: result = function(*args, **kwargs) [ 656.062232] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 656.062232] env[61970]: return func(*args, **kwargs) [ 656.062232] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.062232] env[61970]: raise e [ 656.062232] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.062232] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 656.062232] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.062232] env[61970]: created_port_ids = self._update_ports_for_instance( [ 656.062232] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.062232] env[61970]: with excutils.save_and_reraise_exception(): [ 656.062232] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.062232] env[61970]: self.force_reraise() [ 656.062232] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.062232] env[61970]: raise self.value [ 656.062232] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.062232] env[61970]: updated_port = self._update_port( [ 656.062232] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.062232] env[61970]: _ensure_no_port_binding_failure(port) [ 656.062232] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.062232] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 656.062972] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 6d06502e-9245-444a-8e44-afa3e0c2874b, please check neutron logs for more information. [ 656.062972] env[61970]: Removing descriptor: 17 [ 656.062972] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f2c655-2a31-4bfc-86fd-7960406701e6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.072623] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a7033f-ef81-4cee-ae63-b5d56a165f0d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.091308] env[61970]: ERROR nova.compute.manager [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6d06502e-9245-444a-8e44-afa3e0c2874b, please check neutron logs for more information. [ 656.091308] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Traceback (most recent call last): [ 656.091308] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 656.091308] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] yield resources [ 656.091308] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.091308] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] self.driver.spawn(context, instance, image_meta, [ 656.091308] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 656.091308] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.091308] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.091308] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] vm_ref = self.build_virtual_machine(instance, [ 656.091308] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.091676] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.091676] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.091676] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] for vif in network_info: [ 656.091676] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 656.091676] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] return self._sync_wrapper(fn, *args, **kwargs) [ 656.091676] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 656.091676] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] self.wait() [ 656.091676] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 656.091676] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] self[:] = self._gt.wait() [ 656.091676] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.091676] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] return self._exit_event.wait() [ 656.091676] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 656.091676] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] current.throw(*self._exc) [ 656.092148] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.092148] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] result = function(*args, **kwargs) [ 656.092148] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 656.092148] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] return func(*args, **kwargs) [ 656.092148] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.092148] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] raise e [ 656.092148] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.092148] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] nwinfo = self.network_api.allocate_for_instance( [ 656.092148] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.092148] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] created_port_ids = self._update_ports_for_instance( [ 656.092148] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.092148] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] with excutils.save_and_reraise_exception(): [ 656.092148] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.092548] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] self.force_reraise() [ 656.092548] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.092548] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] raise self.value [ 656.092548] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.092548] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] updated_port = self._update_port( [ 656.092548] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.092548] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] _ensure_no_port_binding_failure(port) [ 656.092548] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.092548] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] raise exception.PortBindingFailed(port_id=port['id']) [ 656.092548] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] nova.exception.PortBindingFailed: Binding failed for port 6d06502e-9245-444a-8e44-afa3e0c2874b, please check neutron logs for more information. [ 656.092548] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] [ 656.092548] env[61970]: INFO nova.compute.manager [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Terminating instance [ 656.093842] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Acquiring lock "refresh_cache-77460934-215e-4294-8eb3-a380dfbacddd" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.269636] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.472400] env[61970]: DEBUG nova.network.neutron [req-f219c466-3fd8-4418-a5bd-50ca05fe9df6 req-81ebd564-55b1-46c2-9df5-063f097ee10f service nova] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.580322] env[61970]: DEBUG nova.network.neutron [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.585535] env[61970]: DEBUG nova.network.neutron [req-f219c466-3fd8-4418-a5bd-50ca05fe9df6 req-81ebd564-55b1-46c2-9df5-063f097ee10f service nova] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.672691] env[61970]: DEBUG nova.network.neutron [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.947017] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52e3e78e-eef9-4903-95eb-0c2526b3fb0e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.954793] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3404ea7-7045-4e3c-bf03-54f0f8906db5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.985303] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a989d4b-364b-4a65-b5e8-9f6202422b1c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.992828] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-383d45a6-9a12-40d0-8b6c-f4158e1c6e65 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.005938] env[61970]: DEBUG nova.compute.provider_tree [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 657.089481] env[61970]: DEBUG oslo_concurrency.lockutils [req-f219c466-3fd8-4418-a5bd-50ca05fe9df6 req-81ebd564-55b1-46c2-9df5-063f097ee10f service nova] Releasing lock "refresh_cache-77460934-215e-4294-8eb3-a380dfbacddd" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.089977] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Acquired lock "refresh_cache-77460934-215e-4294-8eb3-a380dfbacddd" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.090244] env[61970]: DEBUG nova.network.neutron [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 657.176277] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Releasing lock "refresh_cache-816de9f9-a8a9-415a-a6c4-57baea86df51" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.176558] env[61970]: DEBUG nova.compute.manager [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 657.176804] env[61970]: DEBUG nova.compute.manager [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 657.176981] env[61970]: DEBUG nova.network.neutron [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 657.209574] env[61970]: DEBUG nova.network.neutron [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.511027] env[61970]: DEBUG nova.scheduler.client.report [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 657.623551] env[61970]: DEBUG nova.network.neutron [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.713266] env[61970]: DEBUG nova.network.neutron [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.751783] env[61970]: DEBUG nova.network.neutron [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.989065] env[61970]: DEBUG nova.compute.manager [req-176808ad-27b7-4f2e-bc84-b0cdca8072b7 req-cb6d567b-e170-4779-8ef8-499af3386aa7 service nova] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Received event network-vif-deleted-6d06502e-9245-444a-8e44-afa3e0c2874b {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 658.015132] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.970s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.015984] env[61970]: ERROR nova.compute.manager [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f2005d63-aefe-4dc0-b331-c296a931017f, please check neutron logs for more information. [ 658.015984] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Traceback (most recent call last): [ 658.015984] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 658.015984] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] self.driver.spawn(context, instance, image_meta, [ 658.015984] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 658.015984] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] self._vmops.spawn(context, instance, image_meta, injected_files, [ 658.015984] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 658.015984] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] vm_ref = self.build_virtual_machine(instance, [ 658.015984] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 658.015984] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] vif_infos = vmwarevif.get_vif_info(self._session, [ 658.015984] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 658.016287] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] for vif in network_info: [ 658.016287] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 658.016287] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] return self._sync_wrapper(fn, *args, **kwargs) [ 658.016287] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 658.016287] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] self.wait() [ 658.016287] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 658.016287] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] self[:] = self._gt.wait() [ 658.016287] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 658.016287] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] return self._exit_event.wait() [ 658.016287] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 658.016287] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] result = hub.switch() [ 658.016287] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 658.016287] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] return self.greenlet.switch() [ 658.016648] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.016648] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] result = function(*args, **kwargs) [ 658.016648] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 658.016648] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] return func(*args, **kwargs) [ 658.016648] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 658.016648] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] raise e [ 658.016648] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.016648] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] nwinfo = self.network_api.allocate_for_instance( [ 658.016648] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.016648] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] created_port_ids = self._update_ports_for_instance( [ 658.016648] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.016648] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] with excutils.save_and_reraise_exception(): [ 658.016648] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.016950] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] self.force_reraise() [ 658.016950] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.016950] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] raise self.value [ 658.016950] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.016950] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] updated_port = self._update_port( [ 658.016950] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.016950] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] _ensure_no_port_binding_failure(port) [ 658.016950] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.016950] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] raise exception.PortBindingFailed(port_id=port['id']) [ 658.016950] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] nova.exception.PortBindingFailed: Binding failed for port f2005d63-aefe-4dc0-b331-c296a931017f, please check neutron logs for more information. [ 658.016950] env[61970]: ERROR nova.compute.manager [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] [ 658.017220] env[61970]: DEBUG nova.compute.utils [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Binding failed for port f2005d63-aefe-4dc0-b331-c296a931017f, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 658.021289] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.973s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.024252] env[61970]: DEBUG nova.compute.manager [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Build of instance 26f6843a-2548-47e0-9dc9-bc53eda32b60 was re-scheduled: Binding failed for port f2005d63-aefe-4dc0-b331-c296a931017f, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 658.027747] env[61970]: DEBUG nova.compute.manager [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 658.027747] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Acquiring lock "refresh_cache-26f6843a-2548-47e0-9dc9-bc53eda32b60" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.029107] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Acquired lock "refresh_cache-26f6843a-2548-47e0-9dc9-bc53eda32b60" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.029307] env[61970]: DEBUG nova.network.neutron [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 658.214963] env[61970]: INFO nova.compute.manager [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 816de9f9-a8a9-415a-a6c4-57baea86df51] Took 1.04 seconds to deallocate network for instance. [ 658.254434] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Releasing lock "refresh_cache-77460934-215e-4294-8eb3-a380dfbacddd" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.255016] env[61970]: DEBUG nova.compute.manager [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 658.255297] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 658.255684] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-20b9c48a-fbe0-4e31-bcfa-d208371f4dd8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.267967] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e08517-bf8c-4e7c-be3c-d871e17230b0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.294279] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 77460934-215e-4294-8eb3-a380dfbacddd could not be found. [ 658.294482] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 658.294807] env[61970]: INFO nova.compute.manager [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Took 0.04 seconds to destroy the instance on the hypervisor. [ 658.295443] env[61970]: DEBUG oslo.service.loopingcall [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 658.295443] env[61970]: DEBUG nova.compute.manager [-] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.295443] env[61970]: DEBUG nova.network.neutron [-] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 658.318608] env[61970]: DEBUG nova.network.neutron [-] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.439858] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Acquiring lock "95d428e6-fd95-4628-9053-76d218f39258" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.440608] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Lock "95d428e6-fd95-4628-9053-76d218f39258" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.551714] env[61970]: DEBUG nova.network.neutron [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.672033] env[61970]: DEBUG nova.network.neutron [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.824718] env[61970]: DEBUG nova.network.neutron [-] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.022048] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb813c9d-87d6-4ae1-a062-4bf3c310c166 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.030200] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c9560f-6989-4d72-92fe-63a5406ae3a6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.061076] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79303c3b-f763-4601-a3a8-1ea757ec6a73 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.068861] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0378e6dc-b619-44cd-bde1-5a9a1b5a47ba {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.083506] env[61970]: DEBUG nova.compute.provider_tree [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.174353] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Releasing lock "refresh_cache-26f6843a-2548-47e0-9dc9-bc53eda32b60" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.174590] env[61970]: DEBUG nova.compute.manager [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 659.174768] env[61970]: DEBUG nova.compute.manager [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 659.174933] env[61970]: DEBUG nova.network.neutron [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 659.209424] env[61970]: DEBUG nova.network.neutron [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.260629] env[61970]: INFO nova.scheduler.client.report [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Deleted allocations for instance 816de9f9-a8a9-415a-a6c4-57baea86df51 [ 659.329027] env[61970]: INFO nova.compute.manager [-] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Took 1.03 seconds to deallocate network for instance. [ 659.331361] env[61970]: DEBUG nova.compute.claims [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 659.331361] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.593067] env[61970]: DEBUG nova.scheduler.client.report [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.711775] env[61970]: DEBUG nova.network.neutron [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.773131] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0721e599-3e06-475c-911f-a9f087a01be2 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Lock "816de9f9-a8a9-415a-a6c4-57baea86df51" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.815s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.099517] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.078s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.100310] env[61970]: ERROR nova.compute.manager [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 379ab7e2-3794-4982-add9-60278acf0764, please check neutron logs for more information. [ 660.100310] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Traceback (most recent call last): [ 660.100310] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.100310] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] self.driver.spawn(context, instance, image_meta, [ 660.100310] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 660.100310] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.100310] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.100310] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] vm_ref = self.build_virtual_machine(instance, [ 660.100310] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.100310] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.100310] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.100723] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] for vif in network_info: [ 660.100723] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 660.100723] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] return self._sync_wrapper(fn, *args, **kwargs) [ 660.100723] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 660.100723] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] self.wait() [ 660.100723] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 660.100723] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] self[:] = self._gt.wait() [ 660.100723] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.100723] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] return self._exit_event.wait() [ 660.100723] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 660.100723] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] result = hub.switch() [ 660.100723] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 660.100723] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] return self.greenlet.switch() [ 660.101056] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.101056] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] result = function(*args, **kwargs) [ 660.101056] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 660.101056] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] return func(*args, **kwargs) [ 660.101056] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.101056] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] raise e [ 660.101056] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.101056] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] nwinfo = self.network_api.allocate_for_instance( [ 660.101056] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.101056] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] created_port_ids = self._update_ports_for_instance( [ 660.101056] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.101056] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] with excutils.save_and_reraise_exception(): [ 660.101056] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.101388] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] self.force_reraise() [ 660.101388] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.101388] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] raise self.value [ 660.101388] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.101388] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] updated_port = self._update_port( [ 660.101388] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.101388] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] _ensure_no_port_binding_failure(port) [ 660.101388] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.101388] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] raise exception.PortBindingFailed(port_id=port['id']) [ 660.101388] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] nova.exception.PortBindingFailed: Binding failed for port 379ab7e2-3794-4982-add9-60278acf0764, please check neutron logs for more information. [ 660.101388] env[61970]: ERROR nova.compute.manager [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] [ 660.101670] env[61970]: DEBUG nova.compute.utils [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Binding failed for port 379ab7e2-3794-4982-add9-60278acf0764, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 660.102011] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.584s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.103452] env[61970]: INFO nova.compute.claims [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 660.106077] env[61970]: DEBUG nova.compute.manager [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Build of instance 2933b91a-ab88-4f0f-bd63-3752cdc763de was re-scheduled: Binding failed for port 379ab7e2-3794-4982-add9-60278acf0764, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 660.106525] env[61970]: DEBUG nova.compute.manager [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 660.106744] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "refresh_cache-2933b91a-ab88-4f0f-bd63-3752cdc763de" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.106889] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "refresh_cache-2933b91a-ab88-4f0f-bd63-3752cdc763de" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.107058] env[61970]: DEBUG nova.network.neutron [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 660.215446] env[61970]: INFO nova.compute.manager [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] [instance: 26f6843a-2548-47e0-9dc9-bc53eda32b60] Took 1.04 seconds to deallocate network for instance. [ 660.277187] env[61970]: DEBUG nova.compute.manager [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 660.642672] env[61970]: DEBUG nova.network.neutron [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.778326] env[61970]: DEBUG nova.network.neutron [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.804019] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.148126] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Acquiring lock "3f0767a0-9925-4cb9-8032-b5ca071cd572" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.148126] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Lock "3f0767a0-9925-4cb9-8032-b5ca071cd572" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.247032] env[61970]: INFO nova.scheduler.client.report [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Deleted allocations for instance 26f6843a-2548-47e0-9dc9-bc53eda32b60 [ 661.283345] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "refresh_cache-2933b91a-ab88-4f0f-bd63-3752cdc763de" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.283924] env[61970]: DEBUG nova.compute.manager [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 661.283924] env[61970]: DEBUG nova.compute.manager [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 661.283924] env[61970]: DEBUG nova.network.neutron [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 661.305725] env[61970]: DEBUG nova.network.neutron [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.522058] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c19a772-58e0-4933-99b9-9d67082dbcae {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.530341] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e693642-1bfa-4167-ae31-f55c3087bcfc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.565349] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d02d768-fc7e-4f6c-97b3-490c2ceb548d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.575631] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81fdb8d0-ba54-4002-81d1-5a987f3021bb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.592468] env[61970]: DEBUG nova.compute.provider_tree [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.757964] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26db3624-6474-4b51-8c30-fef895a5b8e3 tempest-ServersV294TestFqdnHostnames-250482381 tempest-ServersV294TestFqdnHostnames-250482381-project-member] Lock "26f6843a-2548-47e0-9dc9-bc53eda32b60" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.498s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.809923] env[61970]: DEBUG nova.network.neutron [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.096162] env[61970]: DEBUG nova.scheduler.client.report [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 662.264130] env[61970]: DEBUG nova.compute.manager [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 662.315025] env[61970]: INFO nova.compute.manager [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 2933b91a-ab88-4f0f-bd63-3752cdc763de] Took 1.03 seconds to deallocate network for instance. [ 662.383591] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquiring lock "5ce0137d-a22e-4f22-9d3f-ed4812e67ec7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.383848] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Lock "5ce0137d-a22e-4f22-9d3f-ed4812e67ec7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.411021] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquiring lock "3528d9b6-f3fb-43e9-be27-113d887a3414" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.411021] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Lock "3528d9b6-f3fb-43e9-be27-113d887a3414" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.442235] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquiring lock "07eeb310-312b-4f32-bd5f-7b254cff0fe1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.442487] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Lock "07eeb310-312b-4f32-bd5f-7b254cff0fe1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.603195] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.501s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.603708] env[61970]: DEBUG nova.compute.manager [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 662.606424] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.982s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.793047] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.113579] env[61970]: DEBUG nova.compute.utils [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 663.115150] env[61970]: DEBUG nova.compute.manager [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 663.115325] env[61970]: DEBUG nova.network.neutron [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 663.178962] env[61970]: DEBUG nova.policy [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd04344afddb14418b5018b69fbd7694e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c6a4997482a4834a30c79c4cbab2cc1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 663.343583] env[61970]: INFO nova.scheduler.client.report [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Deleted allocations for instance 2933b91a-ab88-4f0f-bd63-3752cdc763de [ 663.544982] env[61970]: DEBUG nova.network.neutron [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Successfully created port: 9b7f5748-b3b2-41f9-ba64-c6efe0d7f78c {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 663.618591] env[61970]: DEBUG nova.compute.manager [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 663.640993] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance de9fde95-3ad2-414a-b561-fe4ccc1dc9ba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 663.641175] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 293b6863-510c-493e-8c9a-af53ad80fc89 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 663.641301] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 77460934-215e-4294-8eb3-a380dfbacddd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 663.641418] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 90cb79fb-aeb4-4c13-ab80-2787f67ff240 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 663.855635] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9ebb86d9-932e-435a-8456-7d86d1f87adb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "2933b91a-ab88-4f0f-bd63-3752cdc763de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.141s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.148555] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 719de3af-ffe0-4352-9d2c-9e102859fb6c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 664.360178] env[61970]: DEBUG nova.compute.manager [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 664.478999] env[61970]: DEBUG nova.compute.manager [req-15d5c4d1-eb0b-456d-b043-24ce32f82e88 req-79c739a0-b49e-4e35-bc04-73abffb24b0a service nova] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Received event network-changed-9b7f5748-b3b2-41f9-ba64-c6efe0d7f78c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 664.480008] env[61970]: DEBUG nova.compute.manager [req-15d5c4d1-eb0b-456d-b043-24ce32f82e88 req-79c739a0-b49e-4e35-bc04-73abffb24b0a service nova] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Refreshing instance network info cache due to event network-changed-9b7f5748-b3b2-41f9-ba64-c6efe0d7f78c. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 664.480337] env[61970]: DEBUG oslo_concurrency.lockutils [req-15d5c4d1-eb0b-456d-b043-24ce32f82e88 req-79c739a0-b49e-4e35-bc04-73abffb24b0a service nova] Acquiring lock "refresh_cache-90cb79fb-aeb4-4c13-ab80-2787f67ff240" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.480511] env[61970]: DEBUG oslo_concurrency.lockutils [req-15d5c4d1-eb0b-456d-b043-24ce32f82e88 req-79c739a0-b49e-4e35-bc04-73abffb24b0a service nova] Acquired lock "refresh_cache-90cb79fb-aeb4-4c13-ab80-2787f67ff240" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.480716] env[61970]: DEBUG nova.network.neutron [req-15d5c4d1-eb0b-456d-b043-24ce32f82e88 req-79c739a0-b49e-4e35-bc04-73abffb24b0a service nova] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Refreshing network info cache for port 9b7f5748-b3b2-41f9-ba64-c6efe0d7f78c {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 664.589568] env[61970]: ERROR nova.compute.manager [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9b7f5748-b3b2-41f9-ba64-c6efe0d7f78c, please check neutron logs for more information. [ 664.589568] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 664.589568] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.589568] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 664.589568] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.589568] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 664.589568] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.589568] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 664.589568] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.589568] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 664.589568] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.589568] env[61970]: ERROR nova.compute.manager raise self.value [ 664.589568] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.589568] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 664.589568] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.589568] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 664.590230] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.590230] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 664.590230] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9b7f5748-b3b2-41f9-ba64-c6efe0d7f78c, please check neutron logs for more information. [ 664.590230] env[61970]: ERROR nova.compute.manager [ 664.590230] env[61970]: Traceback (most recent call last): [ 664.590230] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 664.590230] env[61970]: listener.cb(fileno) [ 664.590230] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.590230] env[61970]: result = function(*args, **kwargs) [ 664.590230] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 664.590230] env[61970]: return func(*args, **kwargs) [ 664.590230] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.590230] env[61970]: raise e [ 664.590230] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.590230] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 664.590230] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.590230] env[61970]: created_port_ids = self._update_ports_for_instance( [ 664.590230] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.590230] env[61970]: with excutils.save_and_reraise_exception(): [ 664.590230] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.590230] env[61970]: self.force_reraise() [ 664.590230] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.590230] env[61970]: raise self.value [ 664.590230] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.590230] env[61970]: updated_port = self._update_port( [ 664.590230] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.590230] env[61970]: _ensure_no_port_binding_failure(port) [ 664.590230] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.590230] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 664.591282] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 9b7f5748-b3b2-41f9-ba64-c6efe0d7f78c, please check neutron logs for more information. [ 664.591282] env[61970]: Removing descriptor: 15 [ 664.633426] env[61970]: DEBUG nova.compute.manager [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 664.656460] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance b205db9a-8170-4c53-8212-f3fc2ec11d1c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 664.669654] env[61970]: DEBUG nova.virt.hardware [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 664.669947] env[61970]: DEBUG nova.virt.hardware [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 664.670100] env[61970]: DEBUG nova.virt.hardware [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 664.671247] env[61970]: DEBUG nova.virt.hardware [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 664.671247] env[61970]: DEBUG nova.virt.hardware [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 664.671377] env[61970]: DEBUG nova.virt.hardware [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 664.671945] env[61970]: DEBUG nova.virt.hardware [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 664.671945] env[61970]: DEBUG nova.virt.hardware [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 664.671945] env[61970]: DEBUG nova.virt.hardware [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 664.672182] env[61970]: DEBUG nova.virt.hardware [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 664.672285] env[61970]: DEBUG nova.virt.hardware [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 664.673690] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de78be66-88f1-46c2-8560-339a6a942312 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.684019] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2946e683-fb27-48c3-b08b-6bec043d9b1c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.698712] env[61970]: ERROR nova.compute.manager [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9b7f5748-b3b2-41f9-ba64-c6efe0d7f78c, please check neutron logs for more information. [ 664.698712] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Traceback (most recent call last): [ 664.698712] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 664.698712] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] yield resources [ 664.698712] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 664.698712] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] self.driver.spawn(context, instance, image_meta, [ 664.698712] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 664.698712] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] self._vmops.spawn(context, instance, image_meta, injected_files, [ 664.698712] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 664.698712] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] vm_ref = self.build_virtual_machine(instance, [ 664.698712] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 664.699159] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] vif_infos = vmwarevif.get_vif_info(self._session, [ 664.699159] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 664.699159] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] for vif in network_info: [ 664.699159] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 664.699159] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] return self._sync_wrapper(fn, *args, **kwargs) [ 664.699159] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 664.699159] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] self.wait() [ 664.699159] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 664.699159] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] self[:] = self._gt.wait() [ 664.699159] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 664.699159] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] return self._exit_event.wait() [ 664.699159] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 664.699159] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] current.throw(*self._exc) [ 664.699539] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.699539] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] result = function(*args, **kwargs) [ 664.699539] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 664.699539] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] return func(*args, **kwargs) [ 664.699539] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.699539] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] raise e [ 664.699539] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.699539] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] nwinfo = self.network_api.allocate_for_instance( [ 664.699539] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.699539] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] created_port_ids = self._update_ports_for_instance( [ 664.699539] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.699539] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] with excutils.save_and_reraise_exception(): [ 664.699539] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.699872] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] self.force_reraise() [ 664.699872] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.699872] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] raise self.value [ 664.699872] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.699872] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] updated_port = self._update_port( [ 664.699872] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.699872] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] _ensure_no_port_binding_failure(port) [ 664.699872] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.699872] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] raise exception.PortBindingFailed(port_id=port['id']) [ 664.699872] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] nova.exception.PortBindingFailed: Binding failed for port 9b7f5748-b3b2-41f9-ba64-c6efe0d7f78c, please check neutron logs for more information. [ 664.699872] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] [ 664.699872] env[61970]: INFO nova.compute.manager [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Terminating instance [ 664.701561] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "refresh_cache-90cb79fb-aeb4-4c13-ab80-2787f67ff240" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.886921] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.001636] env[61970]: DEBUG nova.network.neutron [req-15d5c4d1-eb0b-456d-b043-24ce32f82e88 req-79c739a0-b49e-4e35-bc04-73abffb24b0a service nova] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.119360] env[61970]: DEBUG nova.network.neutron [req-15d5c4d1-eb0b-456d-b043-24ce32f82e88 req-79c739a0-b49e-4e35-bc04-73abffb24b0a service nova] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.162397] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 665.586429] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "a8ecafbe-f1e1-49f1-945d-ef8b6d15b709" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.586719] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "a8ecafbe-f1e1-49f1-945d-ef8b6d15b709" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.622577] env[61970]: DEBUG oslo_concurrency.lockutils [req-15d5c4d1-eb0b-456d-b043-24ce32f82e88 req-79c739a0-b49e-4e35-bc04-73abffb24b0a service nova] Releasing lock "refresh_cache-90cb79fb-aeb4-4c13-ab80-2787f67ff240" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.623301] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "refresh_cache-90cb79fb-aeb4-4c13-ab80-2787f67ff240" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.623301] env[61970]: DEBUG nova.network.neutron [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 665.665099] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 665.906389] env[61970]: DEBUG oslo_concurrency.lockutils [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.906621] env[61970]: DEBUG oslo_concurrency.lockutils [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.116725] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "a78b63a5-6bb8-4271-90d8-1e86fb29db4f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.116964] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "a78b63a5-6bb8-4271-90d8-1e86fb29db4f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.139687] env[61970]: DEBUG nova.network.neutron [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.167899] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 3661a95f-a75c-4f0f-ad25-428afdb7bccb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.225950] env[61970]: DEBUG nova.network.neutron [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.503384] env[61970]: DEBUG nova.compute.manager [req-2d52950d-9a4b-4bed-94cf-5e876b56389c req-d7413b24-9e27-4853-b325-97240c92a575 service nova] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Received event network-vif-deleted-9b7f5748-b3b2-41f9-ba64-c6efe0d7f78c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 666.671042] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 7dda08a4-83ae-4882-9469-324d948c9749 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.728840] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "refresh_cache-90cb79fb-aeb4-4c13-ab80-2787f67ff240" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.729291] env[61970]: DEBUG nova.compute.manager [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 666.729490] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 666.729802] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf43d80a-6ef4-4c4b-b520-b21b65a73768 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.739184] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a0590bb-2b28-41e2-8c7d-1a810b9abfc2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.760882] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 90cb79fb-aeb4-4c13-ab80-2787f67ff240 could not be found. [ 666.761117] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 666.761305] env[61970]: INFO nova.compute.manager [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Took 0.03 seconds to destroy the instance on the hypervisor. [ 666.761547] env[61970]: DEBUG oslo.service.loopingcall [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 666.762084] env[61970]: DEBUG nova.compute.manager [-] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 666.762183] env[61970]: DEBUG nova.network.neutron [-] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 666.777642] env[61970]: DEBUG nova.network.neutron [-] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.173936] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 9d71be92-36ef-4523-919a-931110d73248 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.280911] env[61970]: DEBUG nova.network.neutron [-] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.678606] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.785537] env[61970]: INFO nova.compute.manager [-] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Took 1.02 seconds to deallocate network for instance. [ 667.791605] env[61970]: DEBUG nova.compute.claims [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 667.791801] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.181860] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 915c3b6c-640b-430f-b264-40bf85c642b9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 668.684810] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 1a9257f6-5493-4ebb-8c37-88e11480a4f6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.187585] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 762efd5e-c95e-4718-98de-2e0b05226c06 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.690256] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance a540095e-563e-4059-b861-5d0e1b4995f4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.193937] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 29ac968a-37db-4c0d-9be9-9f577b53c533 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.697244] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 34719fba-dc16-4b43-ab48-517982ca85a4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 671.201386] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 1dc83f19-2965-418b-8d2a-56aa1124d258 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 671.707043] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance ed10e0a6-c79f-4783-bb93-312c5f0ffd51 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 672.209915] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 5d352f39-4544-46ec-af43-e539d511b47b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 672.713058] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance fd4274ed-bf9a-45ab-a350-a8ebbc5ad404 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 673.215646] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance e74c9188-4f8a-4071-bc4f-c2be91fec3f7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 673.719152] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 01ee68e2-4310-477e-8adb-6ad67d6a65e7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 674.222921] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 95d428e6-fd95-4628-9053-76d218f39258 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 674.726573] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 3f0767a0-9925-4cb9-8032-b5ca071cd572 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 675.229849] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 675.734299] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 3528d9b6-f3fb-43e9-be27-113d887a3414 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 676.238686] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 07eeb310-312b-4f32-bd5f-7b254cff0fe1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 676.238686] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 676.238686] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 676.603662] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8971ffb3-35cc-477a-81e7-b2062f0c4dfe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.610990] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f789b6a6-cb98-47d0-b418-5ab4c5b3d655 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.639183] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ec44c2-9fe5-4298-85d7-5425eab53afc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.646340] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d45371c-7633-4d91-adf3-d76af91c1e9d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.658953] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 677.162170] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.668809] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 677.668809] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 15.062s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.668809] env[61970]: DEBUG oslo_concurrency.lockutils [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.516s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.543170] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e4ec974-ef15-4b21-bae7-de5f639a78c8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.552738] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2216db40-9a77-46ad-85de-47485464bb87 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.585254] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a2431f9-cdf9-49b7-a04a-845cf8518a51 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.592605] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db90c59a-7a07-4898-95d5-9721dca92386 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.605464] env[61970]: DEBUG nova.compute.provider_tree [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.112764] env[61970]: DEBUG nova.scheduler.client.report [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.618325] env[61970]: DEBUG oslo_concurrency.lockutils [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.949s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.620612] env[61970]: ERROR nova.compute.manager [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 58742262-bf56-4023-8e83-5d4ddd05cfcb, please check neutron logs for more information. [ 679.620612] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Traceback (most recent call last): [ 679.620612] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 679.620612] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] self.driver.spawn(context, instance, image_meta, [ 679.620612] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 679.620612] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.620612] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.620612] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] vm_ref = self.build_virtual_machine(instance, [ 679.620612] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.620612] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.620612] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.621066] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] for vif in network_info: [ 679.621066] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 679.621066] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] return self._sync_wrapper(fn, *args, **kwargs) [ 679.621066] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 679.621066] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] self.wait() [ 679.621066] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 679.621066] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] self[:] = self._gt.wait() [ 679.621066] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.621066] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] return self._exit_event.wait() [ 679.621066] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 679.621066] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] current.throw(*self._exc) [ 679.621066] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.621066] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] result = function(*args, **kwargs) [ 679.621428] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 679.621428] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] return func(*args, **kwargs) [ 679.621428] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.621428] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] raise e [ 679.621428] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.621428] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] nwinfo = self.network_api.allocate_for_instance( [ 679.621428] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.621428] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] created_port_ids = self._update_ports_for_instance( [ 679.621428] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.621428] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] with excutils.save_and_reraise_exception(): [ 679.621428] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.621428] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] self.force_reraise() [ 679.621428] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.621786] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] raise self.value [ 679.621786] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.621786] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] updated_port = self._update_port( [ 679.621786] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.621786] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] _ensure_no_port_binding_failure(port) [ 679.621786] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.621786] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] raise exception.PortBindingFailed(port_id=port['id']) [ 679.621786] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] nova.exception.PortBindingFailed: Binding failed for port 58742262-bf56-4023-8e83-5d4ddd05cfcb, please check neutron logs for more information. [ 679.621786] env[61970]: ERROR nova.compute.manager [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] [ 679.621786] env[61970]: DEBUG nova.compute.utils [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Binding failed for port 58742262-bf56-4023-8e83-5d4ddd05cfcb, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 679.622078] env[61970]: DEBUG oslo_concurrency.lockutils [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.762s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.622493] env[61970]: INFO nova.compute.claims [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 679.626397] env[61970]: DEBUG nova.compute.manager [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Build of instance de9fde95-3ad2-414a-b561-fe4ccc1dc9ba was re-scheduled: Binding failed for port 58742262-bf56-4023-8e83-5d4ddd05cfcb, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 679.627029] env[61970]: DEBUG nova.compute.manager [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 679.627159] env[61970]: DEBUG oslo_concurrency.lockutils [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Acquiring lock "refresh_cache-de9fde95-3ad2-414a-b561-fe4ccc1dc9ba" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.627190] env[61970]: DEBUG oslo_concurrency.lockutils [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Acquired lock "refresh_cache-de9fde95-3ad2-414a-b561-fe4ccc1dc9ba" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.627371] env[61970]: DEBUG nova.network.neutron [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 680.147659] env[61970]: DEBUG nova.network.neutron [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.195732] env[61970]: DEBUG nova.network.neutron [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.698840] env[61970]: DEBUG oslo_concurrency.lockutils [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Releasing lock "refresh_cache-de9fde95-3ad2-414a-b561-fe4ccc1dc9ba" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.699182] env[61970]: DEBUG nova.compute.manager [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 680.699300] env[61970]: DEBUG nova.compute.manager [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 680.699482] env[61970]: DEBUG nova.network.neutron [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 680.714470] env[61970]: DEBUG nova.network.neutron [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.066477] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ee4025-557a-4432-babb-6074f209d69c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.074303] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5931f730-8e8f-474b-a83d-f463d51f373f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.102700] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd81cf0c-7e46-4d46-95e1-fdabf7ad7f0c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.109727] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd23db5d-3f0f-4962-9023-49c24de82f81 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.122267] env[61970]: DEBUG nova.compute.provider_tree [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.216985] env[61970]: DEBUG nova.network.neutron [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.625701] env[61970]: DEBUG nova.scheduler.client.report [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 681.719469] env[61970]: INFO nova.compute.manager [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] [instance: de9fde95-3ad2-414a-b561-fe4ccc1dc9ba] Took 1.02 seconds to deallocate network for instance. [ 682.131365] env[61970]: DEBUG oslo_concurrency.lockutils [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.510s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.131365] env[61970]: DEBUG nova.compute.manager [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 682.134642] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.870s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.640275] env[61970]: DEBUG nova.compute.utils [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 682.644400] env[61970]: DEBUG nova.compute.manager [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 682.644575] env[61970]: DEBUG nova.network.neutron [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 682.688596] env[61970]: DEBUG nova.policy [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2af915dc43c84940a937dfb8a04ea0b6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17ca05a1e6664430a02de563d98c1148', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 682.747505] env[61970]: INFO nova.scheduler.client.report [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Deleted allocations for instance de9fde95-3ad2-414a-b561-fe4ccc1dc9ba [ 683.022303] env[61970]: DEBUG nova.network.neutron [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Successfully created port: d00b221d-7660-4f28-a08a-e39ff33fd1ee {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 683.024556] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6983ad1-6d49-4d8e-8bf1-79099952a3c9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.032122] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca1e203-2ee9-497d-aa5e-7f7428a7b5c9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.061103] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73289404-4093-4da6-a2a2-63e97e11effc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.068388] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-861d9652-556b-4108-ae05-a67ca3251e3e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.082645] env[61970]: DEBUG nova.compute.provider_tree [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.148551] env[61970]: DEBUG nova.compute.manager [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 683.257820] env[61970]: DEBUG oslo_concurrency.lockutils [None req-680d3dda-3abd-4dcc-991d-51fe02f6f9e5 tempest-MigrationsAdminTest-1432869150 tempest-MigrationsAdminTest-1432869150-project-member] Lock "de9fde95-3ad2-414a-b561-fe4ccc1dc9ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.868s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.585468] env[61970]: DEBUG nova.scheduler.client.report [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 683.760845] env[61970]: DEBUG nova.compute.manager [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 684.006870] env[61970]: DEBUG nova.compute.manager [req-a4897ba5-9463-4477-bbf3-41cfa536866c req-8f4468f5-ba6b-4716-bc27-1a61f4b90990 service nova] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Received event network-changed-d00b221d-7660-4f28-a08a-e39ff33fd1ee {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 684.007076] env[61970]: DEBUG nova.compute.manager [req-a4897ba5-9463-4477-bbf3-41cfa536866c req-8f4468f5-ba6b-4716-bc27-1a61f4b90990 service nova] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Refreshing instance network info cache due to event network-changed-d00b221d-7660-4f28-a08a-e39ff33fd1ee. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 684.007292] env[61970]: DEBUG oslo_concurrency.lockutils [req-a4897ba5-9463-4477-bbf3-41cfa536866c req-8f4468f5-ba6b-4716-bc27-1a61f4b90990 service nova] Acquiring lock "refresh_cache-719de3af-ffe0-4352-9d2c-9e102859fb6c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.007436] env[61970]: DEBUG oslo_concurrency.lockutils [req-a4897ba5-9463-4477-bbf3-41cfa536866c req-8f4468f5-ba6b-4716-bc27-1a61f4b90990 service nova] Acquired lock "refresh_cache-719de3af-ffe0-4352-9d2c-9e102859fb6c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.007594] env[61970]: DEBUG nova.network.neutron [req-a4897ba5-9463-4477-bbf3-41cfa536866c req-8f4468f5-ba6b-4716-bc27-1a61f4b90990 service nova] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Refreshing network info cache for port d00b221d-7660-4f28-a08a-e39ff33fd1ee {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 684.030792] env[61970]: ERROR nova.compute.manager [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d00b221d-7660-4f28-a08a-e39ff33fd1ee, please check neutron logs for more information. [ 684.030792] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 684.030792] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.030792] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 684.030792] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.030792] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 684.030792] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.030792] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 684.030792] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.030792] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 684.030792] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.030792] env[61970]: ERROR nova.compute.manager raise self.value [ 684.030792] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.030792] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 684.030792] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.030792] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 684.031245] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.031245] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 684.031245] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d00b221d-7660-4f28-a08a-e39ff33fd1ee, please check neutron logs for more information. [ 684.031245] env[61970]: ERROR nova.compute.manager [ 684.031357] env[61970]: Traceback (most recent call last): [ 684.031398] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 684.031398] env[61970]: listener.cb(fileno) [ 684.031398] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.031398] env[61970]: result = function(*args, **kwargs) [ 684.031398] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 684.031398] env[61970]: return func(*args, **kwargs) [ 684.031398] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.031398] env[61970]: raise e [ 684.031658] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.031658] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 684.031658] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.031658] env[61970]: created_port_ids = self._update_ports_for_instance( [ 684.031658] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.031658] env[61970]: with excutils.save_and_reraise_exception(): [ 684.031658] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.031658] env[61970]: self.force_reraise() [ 684.031658] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.031658] env[61970]: raise self.value [ 684.031658] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.031658] env[61970]: updated_port = self._update_port( [ 684.031658] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.031658] env[61970]: _ensure_no_port_binding_failure(port) [ 684.031658] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.031658] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 684.031658] env[61970]: nova.exception.PortBindingFailed: Binding failed for port d00b221d-7660-4f28-a08a-e39ff33fd1ee, please check neutron logs for more information. [ 684.031658] env[61970]: Removing descriptor: 15 [ 684.091478] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.957s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.092122] env[61970]: ERROR nova.compute.manager [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4d3960a5-7523-42c8-9e11-6bfa1e9d8a51, please check neutron logs for more information. [ 684.092122] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Traceback (most recent call last): [ 684.092122] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 684.092122] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] self.driver.spawn(context, instance, image_meta, [ 684.092122] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 684.092122] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.092122] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.092122] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] vm_ref = self.build_virtual_machine(instance, [ 684.092122] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.092122] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.092122] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.092441] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] for vif in network_info: [ 684.092441] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 684.092441] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] return self._sync_wrapper(fn, *args, **kwargs) [ 684.092441] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 684.092441] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] self.wait() [ 684.092441] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 684.092441] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] self[:] = self._gt.wait() [ 684.092441] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.092441] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] return self._exit_event.wait() [ 684.092441] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 684.092441] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] current.throw(*self._exc) [ 684.092441] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.092441] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] result = function(*args, **kwargs) [ 684.092905] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 684.092905] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] return func(*args, **kwargs) [ 684.092905] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.092905] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] raise e [ 684.092905] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.092905] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] nwinfo = self.network_api.allocate_for_instance( [ 684.092905] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.092905] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] created_port_ids = self._update_ports_for_instance( [ 684.092905] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.092905] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] with excutils.save_and_reraise_exception(): [ 684.092905] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.092905] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] self.force_reraise() [ 684.092905] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.093281] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] raise self.value [ 684.093281] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.093281] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] updated_port = self._update_port( [ 684.093281] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.093281] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] _ensure_no_port_binding_failure(port) [ 684.093281] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.093281] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] raise exception.PortBindingFailed(port_id=port['id']) [ 684.093281] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] nova.exception.PortBindingFailed: Binding failed for port 4d3960a5-7523-42c8-9e11-6bfa1e9d8a51, please check neutron logs for more information. [ 684.093281] env[61970]: ERROR nova.compute.manager [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] [ 684.093281] env[61970]: DEBUG nova.compute.utils [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Binding failed for port 4d3960a5-7523-42c8-9e11-6bfa1e9d8a51, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 684.094015] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.573s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.095683] env[61970]: INFO nova.compute.claims [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 684.098274] env[61970]: DEBUG nova.compute.manager [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Build of instance 293b6863-510c-493e-8c9a-af53ad80fc89 was re-scheduled: Binding failed for port 4d3960a5-7523-42c8-9e11-6bfa1e9d8a51, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 684.098708] env[61970]: DEBUG nova.compute.manager [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 684.098930] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Acquiring lock "refresh_cache-293b6863-510c-493e-8c9a-af53ad80fc89" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.099137] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Acquired lock "refresh_cache-293b6863-510c-493e-8c9a-af53ad80fc89" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.099329] env[61970]: DEBUG nova.network.neutron [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 684.162045] env[61970]: DEBUG nova.compute.manager [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 684.191951] env[61970]: DEBUG nova.virt.hardware [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 684.192220] env[61970]: DEBUG nova.virt.hardware [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 684.192380] env[61970]: DEBUG nova.virt.hardware [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 684.192559] env[61970]: DEBUG nova.virt.hardware [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 684.192704] env[61970]: DEBUG nova.virt.hardware [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 684.192855] env[61970]: DEBUG nova.virt.hardware [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 684.193084] env[61970]: DEBUG nova.virt.hardware [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 684.193250] env[61970]: DEBUG nova.virt.hardware [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 684.194572] env[61970]: DEBUG nova.virt.hardware [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 684.194772] env[61970]: DEBUG nova.virt.hardware [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 684.194951] env[61970]: DEBUG nova.virt.hardware [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 684.195837] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a225a097-e3ed-471c-a973-c2dc4d9da211 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.204431] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5994c2a9-2b42-467a-ad0b-78032f4b1d18 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.219524] env[61970]: ERROR nova.compute.manager [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d00b221d-7660-4f28-a08a-e39ff33fd1ee, please check neutron logs for more information. [ 684.219524] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Traceback (most recent call last): [ 684.219524] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 684.219524] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] yield resources [ 684.219524] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 684.219524] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] self.driver.spawn(context, instance, image_meta, [ 684.219524] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 684.219524] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.219524] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.219524] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] vm_ref = self.build_virtual_machine(instance, [ 684.219524] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.219944] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.219944] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.219944] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] for vif in network_info: [ 684.219944] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 684.219944] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] return self._sync_wrapper(fn, *args, **kwargs) [ 684.219944] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 684.219944] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] self.wait() [ 684.219944] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 684.219944] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] self[:] = self._gt.wait() [ 684.219944] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.219944] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] return self._exit_event.wait() [ 684.219944] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 684.219944] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] current.throw(*self._exc) [ 684.220359] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.220359] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] result = function(*args, **kwargs) [ 684.220359] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 684.220359] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] return func(*args, **kwargs) [ 684.220359] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.220359] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] raise e [ 684.220359] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.220359] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] nwinfo = self.network_api.allocate_for_instance( [ 684.220359] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.220359] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] created_port_ids = self._update_ports_for_instance( [ 684.220359] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.220359] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] with excutils.save_and_reraise_exception(): [ 684.220359] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.220748] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] self.force_reraise() [ 684.220748] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.220748] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] raise self.value [ 684.220748] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.220748] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] updated_port = self._update_port( [ 684.220748] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.220748] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] _ensure_no_port_binding_failure(port) [ 684.220748] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.220748] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] raise exception.PortBindingFailed(port_id=port['id']) [ 684.220748] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] nova.exception.PortBindingFailed: Binding failed for port d00b221d-7660-4f28-a08a-e39ff33fd1ee, please check neutron logs for more information. [ 684.220748] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] [ 684.220748] env[61970]: INFO nova.compute.manager [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Terminating instance [ 684.221860] env[61970]: DEBUG oslo_concurrency.lockutils [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "refresh_cache-719de3af-ffe0-4352-9d2c-9e102859fb6c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.289386] env[61970]: DEBUG oslo_concurrency.lockutils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.525138] env[61970]: DEBUG nova.network.neutron [req-a4897ba5-9463-4477-bbf3-41cfa536866c req-8f4468f5-ba6b-4716-bc27-1a61f4b90990 service nova] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.597128] env[61970]: DEBUG nova.network.neutron [req-a4897ba5-9463-4477-bbf3-41cfa536866c req-8f4468f5-ba6b-4716-bc27-1a61f4b90990 service nova] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.622649] env[61970]: DEBUG nova.network.neutron [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.713924] env[61970]: DEBUG nova.network.neutron [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.099593] env[61970]: DEBUG oslo_concurrency.lockutils [req-a4897ba5-9463-4477-bbf3-41cfa536866c req-8f4468f5-ba6b-4716-bc27-1a61f4b90990 service nova] Releasing lock "refresh_cache-719de3af-ffe0-4352-9d2c-9e102859fb6c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.099892] env[61970]: DEBUG oslo_concurrency.lockutils [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "refresh_cache-719de3af-ffe0-4352-9d2c-9e102859fb6c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.100057] env[61970]: DEBUG nova.network.neutron [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 685.215738] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Releasing lock "refresh_cache-293b6863-510c-493e-8c9a-af53ad80fc89" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.215961] env[61970]: DEBUG nova.compute.manager [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 685.216176] env[61970]: DEBUG nova.compute.manager [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 685.216347] env[61970]: DEBUG nova.network.neutron [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 685.240163] env[61970]: DEBUG nova.network.neutron [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.475726] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e001f80-79d2-4193-9227-d5fd76719677 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.483282] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef242e19-6760-41e0-a25c-c7b96e65ce01 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.514765] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa9445c-e95b-434a-91be-d09b88add755 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.523033] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc38f61-78d7-4f31-a44b-2157ba74f912 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.535849] env[61970]: DEBUG nova.compute.provider_tree [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.618120] env[61970]: DEBUG nova.network.neutron [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.742508] env[61970]: DEBUG nova.network.neutron [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.749582] env[61970]: DEBUG nova.network.neutron [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.033109] env[61970]: DEBUG nova.compute.manager [req-b1615ef8-97f9-450f-9e4b-563f69d36ef6 req-70743e4c-919a-4f92-abd5-661769157932 service nova] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Received event network-vif-deleted-d00b221d-7660-4f28-a08a-e39ff33fd1ee {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 686.039245] env[61970]: DEBUG nova.scheduler.client.report [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 686.245531] env[61970]: INFO nova.compute.manager [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] [instance: 293b6863-510c-493e-8c9a-af53ad80fc89] Took 1.03 seconds to deallocate network for instance. [ 686.252172] env[61970]: DEBUG oslo_concurrency.lockutils [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "refresh_cache-719de3af-ffe0-4352-9d2c-9e102859fb6c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.252560] env[61970]: DEBUG nova.compute.manager [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 686.252747] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 686.253031] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-05586430-0950-44aa-a5a1-ea5c6ae133c6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.261917] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f23b336-18c2-4f4f-9080-9d3e602e63fd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.282950] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 719de3af-ffe0-4352-9d2c-9e102859fb6c could not be found. [ 686.283182] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 686.283362] env[61970]: INFO nova.compute.manager [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 686.283610] env[61970]: DEBUG oslo.service.loopingcall [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 686.283833] env[61970]: DEBUG nova.compute.manager [-] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 686.283927] env[61970]: DEBUG nova.network.neutron [-] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 686.304106] env[61970]: DEBUG nova.network.neutron [-] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.543939] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.544492] env[61970]: DEBUG nova.compute.manager [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 686.547291] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.278s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.549921] env[61970]: INFO nova.compute.claims [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 686.806370] env[61970]: DEBUG nova.network.neutron [-] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.054735] env[61970]: DEBUG nova.compute.utils [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 687.058641] env[61970]: DEBUG nova.compute.manager [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 687.058801] env[61970]: DEBUG nova.network.neutron [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 687.116981] env[61970]: DEBUG nova.policy [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ad459b8ec61148669f461418de36f12c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '732e517792c540a5b5a650b04d48dc92', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 687.283686] env[61970]: INFO nova.scheduler.client.report [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Deleted allocations for instance 293b6863-510c-493e-8c9a-af53ad80fc89 [ 687.310484] env[61970]: INFO nova.compute.manager [-] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Took 1.03 seconds to deallocate network for instance. [ 687.313139] env[61970]: DEBUG nova.compute.claims [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 687.313139] env[61970]: DEBUG oslo_concurrency.lockutils [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.550249] env[61970]: DEBUG nova.network.neutron [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Successfully created port: df4144c7-1e44-42cf-a3fe-289f5783650d {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 687.561145] env[61970]: DEBUG nova.compute.manager [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 687.796962] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fe7b5e4-c7e4-40a0-ab6e-4e2a54da7261 tempest-ServerActionsV293TestJSON-415124537 tempest-ServerActionsV293TestJSON-415124537-project-member] Lock "293b6863-510c-493e-8c9a-af53ad80fc89" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.729s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.009639] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-310d1d65-0376-43b9-8d09-2af14e5abbf9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.017278] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d6e933-d2a6-4c09-978a-994c46d3bae7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.049445] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff693b8e-4c9e-4fd1-bf8a-55b7d353b36e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.058320] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d883900-3800-4480-b492-e3060b8875e2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.074432] env[61970]: DEBUG nova.compute.provider_tree [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.304491] env[61970]: DEBUG nova.compute.manager [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 688.578591] env[61970]: DEBUG nova.compute.manager [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 688.581886] env[61970]: DEBUG nova.scheduler.client.report [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 688.619016] env[61970]: DEBUG nova.virt.hardware [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 688.619016] env[61970]: DEBUG nova.virt.hardware [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 688.619016] env[61970]: DEBUG nova.virt.hardware [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 688.619795] env[61970]: DEBUG nova.virt.hardware [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 688.620224] env[61970]: DEBUG nova.virt.hardware [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 688.620504] env[61970]: DEBUG nova.virt.hardware [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 688.620931] env[61970]: DEBUG nova.virt.hardware [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 688.621233] env[61970]: DEBUG nova.virt.hardware [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 688.623035] env[61970]: DEBUG nova.virt.hardware [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 688.623035] env[61970]: DEBUG nova.virt.hardware [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 688.623035] env[61970]: DEBUG nova.virt.hardware [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 688.623035] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d8c8ef-d8a9-454b-a129-72ae5c83500e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.632928] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de0921c9-fe49-4f8c-b562-acd572141d28 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.848105] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.919915] env[61970]: DEBUG nova.compute.manager [req-9b8ac539-0fc7-4293-b70a-141246f3b39b req-5fc9a5ef-41dd-4e05-845e-c0c76efe5395 service nova] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Received event network-changed-df4144c7-1e44-42cf-a3fe-289f5783650d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 688.920138] env[61970]: DEBUG nova.compute.manager [req-9b8ac539-0fc7-4293-b70a-141246f3b39b req-5fc9a5ef-41dd-4e05-845e-c0c76efe5395 service nova] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Refreshing instance network info cache due to event network-changed-df4144c7-1e44-42cf-a3fe-289f5783650d. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 688.920351] env[61970]: DEBUG oslo_concurrency.lockutils [req-9b8ac539-0fc7-4293-b70a-141246f3b39b req-5fc9a5ef-41dd-4e05-845e-c0c76efe5395 service nova] Acquiring lock "refresh_cache-b205db9a-8170-4c53-8212-f3fc2ec11d1c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.920542] env[61970]: DEBUG oslo_concurrency.lockutils [req-9b8ac539-0fc7-4293-b70a-141246f3b39b req-5fc9a5ef-41dd-4e05-845e-c0c76efe5395 service nova] Acquired lock "refresh_cache-b205db9a-8170-4c53-8212-f3fc2ec11d1c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.920636] env[61970]: DEBUG nova.network.neutron [req-9b8ac539-0fc7-4293-b70a-141246f3b39b req-5fc9a5ef-41dd-4e05-845e-c0c76efe5395 service nova] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Refreshing network info cache for port df4144c7-1e44-42cf-a3fe-289f5783650d {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 689.087828] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.540s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.088421] env[61970]: DEBUG nova.compute.manager [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 689.092515] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.761s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.213135] env[61970]: ERROR nova.compute.manager [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port df4144c7-1e44-42cf-a3fe-289f5783650d, please check neutron logs for more information. [ 689.213135] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 689.213135] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.213135] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 689.213135] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.213135] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 689.213135] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.213135] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 689.213135] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.213135] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 689.213135] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.213135] env[61970]: ERROR nova.compute.manager raise self.value [ 689.213135] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.213135] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 689.213135] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.213135] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 689.213727] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.213727] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 689.213727] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port df4144c7-1e44-42cf-a3fe-289f5783650d, please check neutron logs for more information. [ 689.213727] env[61970]: ERROR nova.compute.manager [ 689.213727] env[61970]: Traceback (most recent call last): [ 689.213727] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 689.213727] env[61970]: listener.cb(fileno) [ 689.213727] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.213727] env[61970]: result = function(*args, **kwargs) [ 689.213727] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 689.213727] env[61970]: return func(*args, **kwargs) [ 689.213727] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.213727] env[61970]: raise e [ 689.213727] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.213727] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 689.213727] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.213727] env[61970]: created_port_ids = self._update_ports_for_instance( [ 689.213727] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.213727] env[61970]: with excutils.save_and_reraise_exception(): [ 689.213727] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.213727] env[61970]: self.force_reraise() [ 689.213727] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.213727] env[61970]: raise self.value [ 689.213727] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.213727] env[61970]: updated_port = self._update_port( [ 689.213727] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.213727] env[61970]: _ensure_no_port_binding_failure(port) [ 689.213727] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.213727] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 689.214817] env[61970]: nova.exception.PortBindingFailed: Binding failed for port df4144c7-1e44-42cf-a3fe-289f5783650d, please check neutron logs for more information. [ 689.214817] env[61970]: Removing descriptor: 15 [ 689.214817] env[61970]: ERROR nova.compute.manager [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port df4144c7-1e44-42cf-a3fe-289f5783650d, please check neutron logs for more information. [ 689.214817] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Traceback (most recent call last): [ 689.214817] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 689.214817] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] yield resources [ 689.214817] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 689.214817] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] self.driver.spawn(context, instance, image_meta, [ 689.214817] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 689.214817] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 689.214817] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 689.214817] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] vm_ref = self.build_virtual_machine(instance, [ 689.215172] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 689.215172] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] vif_infos = vmwarevif.get_vif_info(self._session, [ 689.215172] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 689.215172] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] for vif in network_info: [ 689.215172] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 689.215172] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] return self._sync_wrapper(fn, *args, **kwargs) [ 689.215172] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 689.215172] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] self.wait() [ 689.215172] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 689.215172] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] self[:] = self._gt.wait() [ 689.215172] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 689.215172] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] return self._exit_event.wait() [ 689.215172] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 689.215523] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] result = hub.switch() [ 689.215523] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 689.215523] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] return self.greenlet.switch() [ 689.215523] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.215523] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] result = function(*args, **kwargs) [ 689.215523] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 689.215523] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] return func(*args, **kwargs) [ 689.215523] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.215523] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] raise e [ 689.215523] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.215523] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] nwinfo = self.network_api.allocate_for_instance( [ 689.215523] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.215523] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] created_port_ids = self._update_ports_for_instance( [ 689.215882] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.215882] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] with excutils.save_and_reraise_exception(): [ 689.215882] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.215882] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] self.force_reraise() [ 689.215882] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.215882] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] raise self.value [ 689.215882] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.215882] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] updated_port = self._update_port( [ 689.215882] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.215882] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] _ensure_no_port_binding_failure(port) [ 689.215882] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.215882] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] raise exception.PortBindingFailed(port_id=port['id']) [ 689.216486] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] nova.exception.PortBindingFailed: Binding failed for port df4144c7-1e44-42cf-a3fe-289f5783650d, please check neutron logs for more information. [ 689.216486] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] [ 689.216486] env[61970]: INFO nova.compute.manager [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Terminating instance [ 689.217656] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Acquiring lock "refresh_cache-b205db9a-8170-4c53-8212-f3fc2ec11d1c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.450415] env[61970]: DEBUG nova.network.neutron [req-9b8ac539-0fc7-4293-b70a-141246f3b39b req-5fc9a5ef-41dd-4e05-845e-c0c76efe5395 service nova] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.577188] env[61970]: DEBUG nova.network.neutron [req-9b8ac539-0fc7-4293-b70a-141246f3b39b req-5fc9a5ef-41dd-4e05-845e-c0c76efe5395 service nova] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.597111] env[61970]: DEBUG nova.compute.utils [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 689.598490] env[61970]: DEBUG nova.compute.manager [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 689.598658] env[61970]: DEBUG nova.network.neutron [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 689.665225] env[61970]: DEBUG nova.policy [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1cba6a3d457b4b94a7cc9d4899498084', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff88e77a2d634bcba4467ee1d3a1d93a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 690.083039] env[61970]: DEBUG oslo_concurrency.lockutils [req-9b8ac539-0fc7-4293-b70a-141246f3b39b req-5fc9a5ef-41dd-4e05-845e-c0c76efe5395 service nova] Releasing lock "refresh_cache-b205db9a-8170-4c53-8212-f3fc2ec11d1c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.083922] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Acquired lock "refresh_cache-b205db9a-8170-4c53-8212-f3fc2ec11d1c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.083922] env[61970]: DEBUG nova.network.neutron [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 690.104844] env[61970]: DEBUG nova.compute.manager [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 690.107508] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f9e914-17ad-46b9-872d-38572615c0fe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.119029] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3749e461-bced-436b-97a0-8bf2094e1107 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.160800] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-572fb27f-f33d-4757-b44b-3e3e10ac5a85 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.169605] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8422d2-892e-4544-91f2-c5dc6e9fac7e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.188788] env[61970]: DEBUG nova.compute.provider_tree [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.317017] env[61970]: DEBUG nova.network.neutron [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Successfully created port: d332b101-f84b-4534-afc0-f68dbec9462c {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 690.624325] env[61970]: DEBUG nova.network.neutron [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.692029] env[61970]: DEBUG nova.scheduler.client.report [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 690.738147] env[61970]: DEBUG nova.network.neutron [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.961761] env[61970]: DEBUG nova.compute.manager [req-e5620fb6-38d5-4501-a3ab-6bb703cdf81f req-03f1fbf1-ddc2-44db-97b1-3e6439a5ab86 service nova] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Received event network-vif-deleted-df4144c7-1e44-42cf-a3fe-289f5783650d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 691.006724] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Acquiring lock "27996960-d2d4-496c-b52d-5d7312aed040" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.007102] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Lock "27996960-d2d4-496c-b52d-5d7312aed040" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.119287] env[61970]: DEBUG nova.compute.manager [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 691.148095] env[61970]: DEBUG nova.virt.hardware [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 691.148438] env[61970]: DEBUG nova.virt.hardware [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 691.148671] env[61970]: DEBUG nova.virt.hardware [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 691.148792] env[61970]: DEBUG nova.virt.hardware [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 691.148934] env[61970]: DEBUG nova.virt.hardware [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 691.149308] env[61970]: DEBUG nova.virt.hardware [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 691.149382] env[61970]: DEBUG nova.virt.hardware [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 691.150423] env[61970]: DEBUG nova.virt.hardware [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 691.150423] env[61970]: DEBUG nova.virt.hardware [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 691.150423] env[61970]: DEBUG nova.virt.hardware [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 691.150423] env[61970]: DEBUG nova.virt.hardware [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 691.151097] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b37f947-72af-483b-b71b-46064a78f54e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.159168] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54390282-d012-4732-a6c2-65caed9edb5f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.200316] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.108s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.201437] env[61970]: ERROR nova.compute.manager [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6d06502e-9245-444a-8e44-afa3e0c2874b, please check neutron logs for more information. [ 691.201437] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Traceback (most recent call last): [ 691.201437] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 691.201437] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] self.driver.spawn(context, instance, image_meta, [ 691.201437] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 691.201437] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 691.201437] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 691.201437] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] vm_ref = self.build_virtual_machine(instance, [ 691.201437] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 691.201437] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] vif_infos = vmwarevif.get_vif_info(self._session, [ 691.201437] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 691.202439] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] for vif in network_info: [ 691.202439] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 691.202439] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] return self._sync_wrapper(fn, *args, **kwargs) [ 691.202439] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 691.202439] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] self.wait() [ 691.202439] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 691.202439] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] self[:] = self._gt.wait() [ 691.202439] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 691.202439] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] return self._exit_event.wait() [ 691.202439] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 691.202439] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] current.throw(*self._exc) [ 691.202439] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.202439] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] result = function(*args, **kwargs) [ 691.202809] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 691.202809] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] return func(*args, **kwargs) [ 691.202809] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 691.202809] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] raise e [ 691.202809] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.202809] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] nwinfo = self.network_api.allocate_for_instance( [ 691.202809] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.202809] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] created_port_ids = self._update_ports_for_instance( [ 691.202809] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.202809] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] with excutils.save_and_reraise_exception(): [ 691.202809] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.202809] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] self.force_reraise() [ 691.202809] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.203175] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] raise self.value [ 691.203175] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.203175] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] updated_port = self._update_port( [ 691.203175] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.203175] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] _ensure_no_port_binding_failure(port) [ 691.203175] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.203175] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] raise exception.PortBindingFailed(port_id=port['id']) [ 691.203175] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] nova.exception.PortBindingFailed: Binding failed for port 6d06502e-9245-444a-8e44-afa3e0c2874b, please check neutron logs for more information. [ 691.203175] env[61970]: ERROR nova.compute.manager [instance: 77460934-215e-4294-8eb3-a380dfbacddd] [ 691.203175] env[61970]: DEBUG nova.compute.utils [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Binding failed for port 6d06502e-9245-444a-8e44-afa3e0c2874b, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 691.206007] env[61970]: DEBUG nova.compute.manager [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Build of instance 77460934-215e-4294-8eb3-a380dfbacddd was re-scheduled: Binding failed for port 6d06502e-9245-444a-8e44-afa3e0c2874b, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 691.206586] env[61970]: DEBUG nova.compute.manager [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 691.206966] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Acquiring lock "refresh_cache-77460934-215e-4294-8eb3-a380dfbacddd" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.207183] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Acquired lock "refresh_cache-77460934-215e-4294-8eb3-a380dfbacddd" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.207453] env[61970]: DEBUG nova.network.neutron [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 691.211661] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.408s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.211661] env[61970]: INFO nova.compute.claims [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 691.244020] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Releasing lock "refresh_cache-b205db9a-8170-4c53-8212-f3fc2ec11d1c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.244020] env[61970]: DEBUG nova.compute.manager [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 691.244020] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 691.244020] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a2431a2e-99bd-4afb-b9b1-1aff2b1c8111 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.256047] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bcf28ef-1964-4a14-b532-2ccb9bef9557 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.280809] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b205db9a-8170-4c53-8212-f3fc2ec11d1c could not be found. [ 691.280809] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 691.280809] env[61970]: INFO nova.compute.manager [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 691.280809] env[61970]: DEBUG oslo.service.loopingcall [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 691.280809] env[61970]: DEBUG nova.compute.manager [-] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 691.281250] env[61970]: DEBUG nova.network.neutron [-] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 691.303282] env[61970]: DEBUG nova.network.neutron [-] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.753072] env[61970]: DEBUG nova.network.neutron [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.806289] env[61970]: DEBUG nova.network.neutron [-] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.833729] env[61970]: ERROR nova.compute.manager [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d332b101-f84b-4534-afc0-f68dbec9462c, please check neutron logs for more information. [ 691.833729] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 691.833729] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.833729] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 691.833729] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.833729] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 691.833729] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.833729] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 691.833729] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.833729] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 691.833729] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.833729] env[61970]: ERROR nova.compute.manager raise self.value [ 691.833729] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.833729] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 691.833729] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.833729] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 691.834265] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.834265] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 691.834265] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d332b101-f84b-4534-afc0-f68dbec9462c, please check neutron logs for more information. [ 691.834265] env[61970]: ERROR nova.compute.manager [ 691.834265] env[61970]: Traceback (most recent call last): [ 691.834265] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 691.834265] env[61970]: listener.cb(fileno) [ 691.834265] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.834265] env[61970]: result = function(*args, **kwargs) [ 691.834265] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 691.834265] env[61970]: return func(*args, **kwargs) [ 691.834265] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 691.834265] env[61970]: raise e [ 691.834265] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.834265] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 691.834265] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.834265] env[61970]: created_port_ids = self._update_ports_for_instance( [ 691.834265] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.834265] env[61970]: with excutils.save_and_reraise_exception(): [ 691.834265] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.834265] env[61970]: self.force_reraise() [ 691.834265] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.834265] env[61970]: raise self.value [ 691.834265] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.834265] env[61970]: updated_port = self._update_port( [ 691.834265] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.834265] env[61970]: _ensure_no_port_binding_failure(port) [ 691.834265] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.834265] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 691.834944] env[61970]: nova.exception.PortBindingFailed: Binding failed for port d332b101-f84b-4534-afc0-f68dbec9462c, please check neutron logs for more information. [ 691.834944] env[61970]: Removing descriptor: 15 [ 691.834944] env[61970]: ERROR nova.compute.manager [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d332b101-f84b-4534-afc0-f68dbec9462c, please check neutron logs for more information. [ 691.834944] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Traceback (most recent call last): [ 691.834944] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 691.834944] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] yield resources [ 691.834944] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 691.834944] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] self.driver.spawn(context, instance, image_meta, [ 691.834944] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 691.834944] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 691.834944] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 691.834944] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] vm_ref = self.build_virtual_machine(instance, [ 691.835307] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 691.835307] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] vif_infos = vmwarevif.get_vif_info(self._session, [ 691.835307] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 691.835307] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] for vif in network_info: [ 691.835307] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 691.835307] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] return self._sync_wrapper(fn, *args, **kwargs) [ 691.835307] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 691.835307] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] self.wait() [ 691.835307] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 691.835307] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] self[:] = self._gt.wait() [ 691.835307] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 691.835307] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] return self._exit_event.wait() [ 691.835307] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 691.835637] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] result = hub.switch() [ 691.835637] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 691.835637] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] return self.greenlet.switch() [ 691.835637] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.835637] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] result = function(*args, **kwargs) [ 691.835637] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 691.835637] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] return func(*args, **kwargs) [ 691.835637] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 691.835637] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] raise e [ 691.835637] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.835637] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] nwinfo = self.network_api.allocate_for_instance( [ 691.835637] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.835637] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] created_port_ids = self._update_ports_for_instance( [ 691.835957] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.835957] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] with excutils.save_and_reraise_exception(): [ 691.835957] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.835957] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] self.force_reraise() [ 691.835957] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.835957] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] raise self.value [ 691.835957] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.835957] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] updated_port = self._update_port( [ 691.835957] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.835957] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] _ensure_no_port_binding_failure(port) [ 691.835957] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.835957] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] raise exception.PortBindingFailed(port_id=port['id']) [ 691.836317] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] nova.exception.PortBindingFailed: Binding failed for port d332b101-f84b-4534-afc0-f68dbec9462c, please check neutron logs for more information. [ 691.836317] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] [ 691.836317] env[61970]: INFO nova.compute.manager [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Terminating instance [ 691.839107] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Acquiring lock "refresh_cache-84c6fdb9-0858-4f25-bdc8-474ffb9f6abb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.839271] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Acquired lock "refresh_cache-84c6fdb9-0858-4f25-bdc8-474ffb9f6abb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.839533] env[61970]: DEBUG nova.network.neutron [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 691.929100] env[61970]: DEBUG nova.network.neutron [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.313746] env[61970]: INFO nova.compute.manager [-] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Took 1.03 seconds to deallocate network for instance. [ 692.315858] env[61970]: DEBUG nova.compute.claims [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 692.316043] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.361490] env[61970]: DEBUG nova.network.neutron [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.437189] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Releasing lock "refresh_cache-77460934-215e-4294-8eb3-a380dfbacddd" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.437418] env[61970]: DEBUG nova.compute.manager [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 692.437631] env[61970]: DEBUG nova.compute.manager [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 692.437810] env[61970]: DEBUG nova.network.neutron [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 692.470137] env[61970]: DEBUG nova.network.neutron [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.537042] env[61970]: DEBUG nova.network.neutron [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.611671] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21329551-6d8e-4ba1-bf2f-aa657d138079 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.618999] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a44c9388-e49e-4bce-a399-e8117db7434e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.654958] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a02557-c31f-40d2-be25-cdfeafddaefa {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.662386] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f5d65a-f41f-4538-b6f3-b3290ee7b483 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.676478] env[61970]: DEBUG nova.compute.provider_tree [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.973652] env[61970]: DEBUG nova.network.neutron [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.029576] env[61970]: DEBUG nova.compute.manager [req-5421788e-e681-4828-ba72-ec84de47a04b req-3bfc8960-3f31-4fe3-9d91-fc951894d38b service nova] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Received event network-changed-d332b101-f84b-4534-afc0-f68dbec9462c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 693.030015] env[61970]: DEBUG nova.compute.manager [req-5421788e-e681-4828-ba72-ec84de47a04b req-3bfc8960-3f31-4fe3-9d91-fc951894d38b service nova] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Refreshing instance network info cache due to event network-changed-d332b101-f84b-4534-afc0-f68dbec9462c. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 693.030964] env[61970]: DEBUG oslo_concurrency.lockutils [req-5421788e-e681-4828-ba72-ec84de47a04b req-3bfc8960-3f31-4fe3-9d91-fc951894d38b service nova] Acquiring lock "refresh_cache-84c6fdb9-0858-4f25-bdc8-474ffb9f6abb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.040910] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Releasing lock "refresh_cache-84c6fdb9-0858-4f25-bdc8-474ffb9f6abb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.041265] env[61970]: DEBUG nova.compute.manager [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 693.041450] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 693.041840] env[61970]: DEBUG oslo_concurrency.lockutils [req-5421788e-e681-4828-ba72-ec84de47a04b req-3bfc8960-3f31-4fe3-9d91-fc951894d38b service nova] Acquired lock "refresh_cache-84c6fdb9-0858-4f25-bdc8-474ffb9f6abb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.041927] env[61970]: DEBUG nova.network.neutron [req-5421788e-e681-4828-ba72-ec84de47a04b req-3bfc8960-3f31-4fe3-9d91-fc951894d38b service nova] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Refreshing network info cache for port d332b101-f84b-4534-afc0-f68dbec9462c {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 693.043172] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f067abb4-f9ff-4989-815a-8710c6b357da {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.055929] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47b19c0-400f-47da-9fd3-4539e8cb6f16 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.078966] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb could not be found. [ 693.079470] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 693.079687] env[61970]: INFO nova.compute.manager [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Took 0.04 seconds to destroy the instance on the hypervisor. [ 693.080142] env[61970]: DEBUG oslo.service.loopingcall [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 693.080379] env[61970]: DEBUG nova.compute.manager [-] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 693.080471] env[61970]: DEBUG nova.network.neutron [-] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 693.111112] env[61970]: DEBUG nova.network.neutron [-] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.179433] env[61970]: DEBUG nova.scheduler.client.report [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 693.479688] env[61970]: INFO nova.compute.manager [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] [instance: 77460934-215e-4294-8eb3-a380dfbacddd] Took 1.04 seconds to deallocate network for instance. [ 693.573202] env[61970]: DEBUG nova.network.neutron [req-5421788e-e681-4828-ba72-ec84de47a04b req-3bfc8960-3f31-4fe3-9d91-fc951894d38b service nova] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.614605] env[61970]: DEBUG nova.network.neutron [-] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.690359] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.481s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.691105] env[61970]: DEBUG nova.compute.manager [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 693.697479] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.905s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.699469] env[61970]: INFO nova.compute.claims [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 693.886816] env[61970]: DEBUG nova.network.neutron [req-5421788e-e681-4828-ba72-ec84de47a04b req-3bfc8960-3f31-4fe3-9d91-fc951894d38b service nova] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.117041] env[61970]: INFO nova.compute.manager [-] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Took 1.04 seconds to deallocate network for instance. [ 694.119592] env[61970]: DEBUG nova.compute.claims [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 694.119808] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.199014] env[61970]: DEBUG nova.compute.utils [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 694.200523] env[61970]: DEBUG nova.compute.manager [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Not allocating networking since 'none' was specified. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 694.395809] env[61970]: DEBUG oslo_concurrency.lockutils [req-5421788e-e681-4828-ba72-ec84de47a04b req-3bfc8960-3f31-4fe3-9d91-fc951894d38b service nova] Releasing lock "refresh_cache-84c6fdb9-0858-4f25-bdc8-474ffb9f6abb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.396440] env[61970]: DEBUG nova.compute.manager [req-5421788e-e681-4828-ba72-ec84de47a04b req-3bfc8960-3f31-4fe3-9d91-fc951894d38b service nova] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Received event network-vif-deleted-d332b101-f84b-4534-afc0-f68dbec9462c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 694.523044] env[61970]: INFO nova.scheduler.client.report [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Deleted allocations for instance 77460934-215e-4294-8eb3-a380dfbacddd [ 694.702908] env[61970]: DEBUG nova.compute.manager [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 695.042532] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2e80a9-1b81-49f7-a3c4-fb58245fca2a tempest-InstanceActionsTestJSON-1175147553 tempest-InstanceActionsTestJSON-1175147553-project-member] Lock "77460934-215e-4294-8eb3-a380dfbacddd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.290s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.113153] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Acquiring lock "c5c253b6-ebab-4fea-bcc4-f3bb3960070b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.113516] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Lock "c5c253b6-ebab-4fea-bcc4-f3bb3960070b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.196585] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f22f84-a6cf-4fb4-9a60-1ce6d79a1b76 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.205604] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088ae7ec-f682-4093-b606-938cfa2f4f19 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.240140] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b9fa4c-89e3-4d88-b24d-1de37f9424a8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.248667] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be3c15f6-f791-428e-a930-fc4823f9e468 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.263499] env[61970]: DEBUG nova.compute.provider_tree [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.549377] env[61970]: DEBUG nova.compute.manager [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 695.716703] env[61970]: DEBUG nova.compute.manager [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 695.754546] env[61970]: DEBUG nova.virt.hardware [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 695.754791] env[61970]: DEBUG nova.virt.hardware [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 695.754944] env[61970]: DEBUG nova.virt.hardware [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 695.755170] env[61970]: DEBUG nova.virt.hardware [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 695.755435] env[61970]: DEBUG nova.virt.hardware [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 695.755563] env[61970]: DEBUG nova.virt.hardware [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 695.755782] env[61970]: DEBUG nova.virt.hardware [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 695.755938] env[61970]: DEBUG nova.virt.hardware [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 695.757251] env[61970]: DEBUG nova.virt.hardware [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 695.757251] env[61970]: DEBUG nova.virt.hardware [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 695.757251] env[61970]: DEBUG nova.virt.hardware [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 695.757466] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c587303-6c23-4c8f-9f91-4888ef85c480 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.765647] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1496bfc-e137-4a29-a828-b813577562b5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.769890] env[61970]: DEBUG nova.scheduler.client.report [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 695.782672] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Instance VIF info [] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 695.788474] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Creating folder: Project (8281a35cf0ea42a398589d4093b56bf0). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 695.789783] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-881dc784-f6dc-4c97-bd4b-8ded17331617 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.800084] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Created folder: Project (8281a35cf0ea42a398589d4093b56bf0) in parent group-v288740. [ 695.800350] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Creating folder: Instances. Parent ref: group-v288752. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 695.800491] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cbaf9f32-b088-48a8-b128-7b79721fcd1a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.808453] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Created folder: Instances in parent group-v288752. [ 695.808699] env[61970]: DEBUG oslo.service.loopingcall [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 695.808884] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 695.809088] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-143bfc6c-b036-462f-be76-e6cc8ab6eafd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.825512] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 695.825512] env[61970]: value = "task-1355501" [ 695.825512] env[61970]: _type = "Task" [ 695.825512] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.832881] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355501, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.073487] env[61970]: DEBUG oslo_concurrency.lockutils [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.277062] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.579s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.277564] env[61970]: DEBUG nova.compute.manager [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 696.280295] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.394s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.281814] env[61970]: INFO nova.compute.claims [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 696.336857] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355501, 'name': CreateVM_Task, 'duration_secs': 0.324821} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.337497] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 696.337944] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.338124] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.338606] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 696.338858] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74da9492-4c1d-49ff-a743-18fb1701bb1e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.343588] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 696.343588] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52c1f821-f9fb-1571-8c8a-23640b0cf5b8" [ 696.343588] env[61970]: _type = "Task" [ 696.343588] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.351152] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52c1f821-f9fb-1571-8c8a-23640b0cf5b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.786248] env[61970]: DEBUG nova.compute.utils [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 696.790171] env[61970]: DEBUG nova.compute.manager [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 696.790171] env[61970]: DEBUG nova.network.neutron [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 696.835214] env[61970]: DEBUG nova.policy [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd86a91a5e0a94fb38ebfe674de51d9fd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6006519fd046405c950b4f3e9c8f8103', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 696.853960] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52c1f821-f9fb-1571-8c8a-23640b0cf5b8, 'name': SearchDatastore_Task, 'duration_secs': 0.00945} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.854353] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.855031] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 696.855031] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.855031] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.855031] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 696.855263] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9420ffba-7dc6-4f1b-a6b6-267d25b98232 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.862924] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 696.863118] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 696.863796] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8a7d975-c758-4b11-8062-7644efebb099 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.869511] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 696.869511] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5206f4ff-a37e-46c3-656a-778c9c566f19" [ 696.869511] env[61970]: _type = "Task" [ 696.869511] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.880834] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5206f4ff-a37e-46c3-656a-778c9c566f19, 'name': SearchDatastore_Task, 'duration_secs': 0.007276} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.881526] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39d47146-e548-4a8f-82db-51fc499b4e0e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.886563] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 696.886563] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5277f991-e9b1-edf7-68b8-6be6b49ba779" [ 696.886563] env[61970]: _type = "Task" [ 696.886563] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.894133] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5277f991-e9b1-edf7-68b8-6be6b49ba779, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.208865] env[61970]: DEBUG nova.network.neutron [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Successfully created port: 3639de38-74f5-4cd0-ae6e-223790e16f20 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 697.298826] env[61970]: DEBUG nova.compute.manager [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 697.398134] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5277f991-e9b1-edf7-68b8-6be6b49ba779, 'name': SearchDatastore_Task, 'duration_secs': 0.007526} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.398447] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.398793] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8/9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 697.399190] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-663ca89e-743a-4c8f-8a03-38468729905f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.409203] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 697.409203] env[61970]: value = "task-1355502" [ 697.409203] env[61970]: _type = "Task" [ 697.409203] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.417984] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355502, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.797447] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8edf5401-bc6f-43bb-b7fe-ced34fa105be {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.805822] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da46e9a-e205-4d36-8118-542fb28061af {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.850811] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ccaeb10-915b-4aa6-b819-75ed28b9f402 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.860044] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a5f65ea-fad4-4022-b287-53acec861615 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.876810] env[61970]: DEBUG nova.compute.provider_tree [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 697.924794] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355502, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478175} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.925057] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8/9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 697.925273] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 697.925518] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b85ef59a-15af-4a72-b814-9b52c3f468f5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.934691] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 697.934691] env[61970]: value = "task-1355503" [ 697.934691] env[61970]: _type = "Task" [ 697.934691] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.939169] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355503, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.118619] env[61970]: DEBUG nova.compute.manager [req-18e99857-a032-432b-8a8c-fe268bb31562 req-f276f803-dace-4b18-9768-707fb0a3bab4 service nova] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Received event network-changed-3639de38-74f5-4cd0-ae6e-223790e16f20 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 698.118710] env[61970]: DEBUG nova.compute.manager [req-18e99857-a032-432b-8a8c-fe268bb31562 req-f276f803-dace-4b18-9768-707fb0a3bab4 service nova] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Refreshing instance network info cache due to event network-changed-3639de38-74f5-4cd0-ae6e-223790e16f20. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 698.118915] env[61970]: DEBUG oslo_concurrency.lockutils [req-18e99857-a032-432b-8a8c-fe268bb31562 req-f276f803-dace-4b18-9768-707fb0a3bab4 service nova] Acquiring lock "refresh_cache-3661a95f-a75c-4f0f-ad25-428afdb7bccb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.119165] env[61970]: DEBUG oslo_concurrency.lockutils [req-18e99857-a032-432b-8a8c-fe268bb31562 req-f276f803-dace-4b18-9768-707fb0a3bab4 service nova] Acquired lock "refresh_cache-3661a95f-a75c-4f0f-ad25-428afdb7bccb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.119298] env[61970]: DEBUG nova.network.neutron [req-18e99857-a032-432b-8a8c-fe268bb31562 req-f276f803-dace-4b18-9768-707fb0a3bab4 service nova] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Refreshing network info cache for port 3639de38-74f5-4cd0-ae6e-223790e16f20 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 698.289579] env[61970]: ERROR nova.compute.manager [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3639de38-74f5-4cd0-ae6e-223790e16f20, please check neutron logs for more information. [ 698.289579] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 698.289579] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 698.289579] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 698.289579] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.289579] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 698.289579] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.289579] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 698.289579] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.289579] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 698.289579] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.289579] env[61970]: ERROR nova.compute.manager raise self.value [ 698.289579] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.289579] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 698.289579] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.289579] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 698.290596] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.290596] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 698.290596] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3639de38-74f5-4cd0-ae6e-223790e16f20, please check neutron logs for more information. [ 698.290596] env[61970]: ERROR nova.compute.manager [ 698.290596] env[61970]: Traceback (most recent call last): [ 698.290596] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 698.290596] env[61970]: listener.cb(fileno) [ 698.290596] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 698.290596] env[61970]: result = function(*args, **kwargs) [ 698.290596] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 698.290596] env[61970]: return func(*args, **kwargs) [ 698.290596] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 698.290596] env[61970]: raise e [ 698.290596] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 698.290596] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 698.290596] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.290596] env[61970]: created_port_ids = self._update_ports_for_instance( [ 698.290596] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.290596] env[61970]: with excutils.save_and_reraise_exception(): [ 698.290596] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.290596] env[61970]: self.force_reraise() [ 698.290596] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.290596] env[61970]: raise self.value [ 698.290596] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.290596] env[61970]: updated_port = self._update_port( [ 698.290596] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.290596] env[61970]: _ensure_no_port_binding_failure(port) [ 698.290596] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.290596] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 698.291330] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 3639de38-74f5-4cd0-ae6e-223790e16f20, please check neutron logs for more information. [ 698.291330] env[61970]: Removing descriptor: 17 [ 698.316206] env[61970]: DEBUG nova.compute.manager [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 698.340596] env[61970]: DEBUG nova.virt.hardware [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 698.340792] env[61970]: DEBUG nova.virt.hardware [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 698.340864] env[61970]: DEBUG nova.virt.hardware [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 698.341152] env[61970]: DEBUG nova.virt.hardware [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 698.341283] env[61970]: DEBUG nova.virt.hardware [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 698.341369] env[61970]: DEBUG nova.virt.hardware [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 698.341541] env[61970]: DEBUG nova.virt.hardware [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 698.341718] env[61970]: DEBUG nova.virt.hardware [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 698.341916] env[61970]: DEBUG nova.virt.hardware [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 698.342097] env[61970]: DEBUG nova.virt.hardware [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 698.342281] env[61970]: DEBUG nova.virt.hardware [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 698.343174] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a01f89-ada9-4e94-88f1-1cb2e91d4954 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.351273] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bfbc960-5161-4cf1-a2be-7ab73bbbd906 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.365399] env[61970]: ERROR nova.compute.manager [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3639de38-74f5-4cd0-ae6e-223790e16f20, please check neutron logs for more information. [ 698.365399] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Traceback (most recent call last): [ 698.365399] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 698.365399] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] yield resources [ 698.365399] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 698.365399] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] self.driver.spawn(context, instance, image_meta, [ 698.365399] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 698.365399] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 698.365399] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 698.365399] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] vm_ref = self.build_virtual_machine(instance, [ 698.365399] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 698.365964] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] vif_infos = vmwarevif.get_vif_info(self._session, [ 698.365964] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 698.365964] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] for vif in network_info: [ 698.365964] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 698.365964] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] return self._sync_wrapper(fn, *args, **kwargs) [ 698.365964] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 698.365964] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] self.wait() [ 698.365964] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 698.365964] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] self[:] = self._gt.wait() [ 698.365964] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 698.365964] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] return self._exit_event.wait() [ 698.365964] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 698.365964] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] current.throw(*self._exc) [ 698.366594] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 698.366594] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] result = function(*args, **kwargs) [ 698.366594] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 698.366594] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] return func(*args, **kwargs) [ 698.366594] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 698.366594] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] raise e [ 698.366594] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 698.366594] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] nwinfo = self.network_api.allocate_for_instance( [ 698.366594] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.366594] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] created_port_ids = self._update_ports_for_instance( [ 698.366594] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.366594] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] with excutils.save_and_reraise_exception(): [ 698.366594] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.367243] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] self.force_reraise() [ 698.367243] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.367243] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] raise self.value [ 698.367243] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.367243] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] updated_port = self._update_port( [ 698.367243] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.367243] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] _ensure_no_port_binding_failure(port) [ 698.367243] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.367243] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] raise exception.PortBindingFailed(port_id=port['id']) [ 698.367243] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] nova.exception.PortBindingFailed: Binding failed for port 3639de38-74f5-4cd0-ae6e-223790e16f20, please check neutron logs for more information. [ 698.367243] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] [ 698.367243] env[61970]: INFO nova.compute.manager [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Terminating instance [ 698.367938] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Acquiring lock "refresh_cache-3661a95f-a75c-4f0f-ad25-428afdb7bccb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.379520] env[61970]: DEBUG nova.scheduler.client.report [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 698.441819] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355503, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.10016} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.442289] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 698.442863] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd9df2d9-b0ee-4963-9029-513078927e93 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.462444] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Reconfiguring VM instance instance-00000022 to attach disk [datastore2] 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8/9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 698.462765] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66ff3aad-d4b3-4dd8-8d8a-4d4d5fa294ed {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.482619] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 698.482619] env[61970]: value = "task-1355504" [ 698.482619] env[61970]: _type = "Task" [ 698.482619] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.490613] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355504, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.641717] env[61970]: DEBUG nova.network.neutron [req-18e99857-a032-432b-8a8c-fe268bb31562 req-f276f803-dace-4b18-9768-707fb0a3bab4 service nova] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.757907] env[61970]: DEBUG nova.network.neutron [req-18e99857-a032-432b-8a8c-fe268bb31562 req-f276f803-dace-4b18-9768-707fb0a3bab4 service nova] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.884854] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.604s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.885286] env[61970]: DEBUG nova.compute.manager [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 698.888257] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.096s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.994195] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355504, 'name': ReconfigVM_Task, 'duration_secs': 0.259604} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.994741] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Reconfigured VM instance instance-00000022 to attach disk [datastore2] 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8/9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 698.995042] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e8d9c7ec-2b0e-4ecb-afab-5dfb34d55bc1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.003224] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 699.003224] env[61970]: value = "task-1355505" [ 699.003224] env[61970]: _type = "Task" [ 699.003224] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.011479] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355505, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.260270] env[61970]: DEBUG oslo_concurrency.lockutils [req-18e99857-a032-432b-8a8c-fe268bb31562 req-f276f803-dace-4b18-9768-707fb0a3bab4 service nova] Releasing lock "refresh_cache-3661a95f-a75c-4f0f-ad25-428afdb7bccb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.260682] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Acquired lock "refresh_cache-3661a95f-a75c-4f0f-ad25-428afdb7bccb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.261848] env[61970]: DEBUG nova.network.neutron [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 699.392511] env[61970]: DEBUG nova.compute.utils [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 699.393868] env[61970]: DEBUG nova.compute.manager [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 699.394720] env[61970]: DEBUG nova.network.neutron [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 699.462510] env[61970]: DEBUG nova.policy [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7942b1ab4d97493daf6bfef8912f2ddc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'db143ea38da943d29517d4f598ef7a95', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 699.491156] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "5d4d97e1-5812-4884-a1ea-6eb0cee591c0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.491525] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "5d4d97e1-5812-4884-a1ea-6eb0cee591c0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.513770] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355505, 'name': Rename_Task, 'duration_secs': 0.177578} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.516038] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 699.516038] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-728ac1a5-3534-42e2-8fee-cd24e5f4bcf6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.521028] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 699.521028] env[61970]: value = "task-1355506" [ 699.521028] env[61970]: _type = "Task" [ 699.521028] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.533786] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355506, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.720858] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "6345d239-fbae-4f3a-9f61-79e10539ff0b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.721054] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "6345d239-fbae-4f3a-9f61-79e10539ff0b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.786203] env[61970]: DEBUG nova.network.neutron [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.890823] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5785c3-d26a-4c47-8338-1f2860560b43 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.899753] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f99168-5970-4c6f-8737-dc30cc07461b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.903358] env[61970]: DEBUG nova.compute.manager [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 699.937505] env[61970]: DEBUG nova.network.neutron [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Successfully created port: ce17d99a-8184-48b2-a9ac-2afbf5f36def {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 699.940198] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d307e294-762d-4541-8a34-adf3c17d592e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.944420] env[61970]: DEBUG nova.network.neutron [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.949421] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a70b48f-bc56-4860-aa98-7516a2f06666 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.969327] env[61970]: DEBUG nova.compute.provider_tree [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.031891] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355506, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.073263] env[61970]: DEBUG oslo_concurrency.lockutils [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "e2f51299-56fe-46cc-9e08-2246178b9db7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.073468] env[61970]: DEBUG oslo_concurrency.lockutils [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "e2f51299-56fe-46cc-9e08-2246178b9db7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.145616] env[61970]: DEBUG nova.compute.manager [req-73c0c0ab-399a-4ef4-b387-333506b51526 req-966508be-4a47-4705-8eb9-f4eb089193ce service nova] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Received event network-vif-deleted-3639de38-74f5-4cd0-ae6e-223790e16f20 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 700.451017] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Releasing lock "refresh_cache-3661a95f-a75c-4f0f-ad25-428afdb7bccb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.451017] env[61970]: DEBUG nova.compute.manager [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 700.451017] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 700.451017] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fc2c8a5a-d10b-45e9-9ec2-7d7d5b2bbe99 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.465196] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa942e4-50ad-41f7-ad27-d9653ce380f5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.477246] env[61970]: DEBUG nova.scheduler.client.report [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 700.492867] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3661a95f-a75c-4f0f-ad25-428afdb7bccb could not be found. [ 700.492867] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 700.492867] env[61970]: INFO nova.compute.manager [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Took 0.04 seconds to destroy the instance on the hypervisor. [ 700.493133] env[61970]: DEBUG oslo.service.loopingcall [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 700.493649] env[61970]: DEBUG nova.compute.manager [-] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 700.493749] env[61970]: DEBUG nova.network.neutron [-] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 700.513264] env[61970]: DEBUG nova.network.neutron [-] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.531924] env[61970]: DEBUG oslo_vmware.api [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355506, 'name': PowerOnVM_Task, 'duration_secs': 0.569043} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.532062] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 700.532261] env[61970]: INFO nova.compute.manager [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Took 4.82 seconds to spawn the instance on the hypervisor. [ 700.532440] env[61970]: DEBUG nova.compute.manager [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 700.533225] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9299cba-9023-45c8-92cb-4413d72cf162 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.915293] env[61970]: DEBUG nova.compute.manager [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 700.945127] env[61970]: DEBUG nova.virt.hardware [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 700.945381] env[61970]: DEBUG nova.virt.hardware [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 700.945531] env[61970]: DEBUG nova.virt.hardware [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 700.945867] env[61970]: DEBUG nova.virt.hardware [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 700.945867] env[61970]: DEBUG nova.virt.hardware [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 700.945958] env[61970]: DEBUG nova.virt.hardware [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 700.948272] env[61970]: DEBUG nova.virt.hardware [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 700.948453] env[61970]: DEBUG nova.virt.hardware [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 700.948687] env[61970]: DEBUG nova.virt.hardware [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 700.948865] env[61970]: DEBUG nova.virt.hardware [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 700.949120] env[61970]: DEBUG nova.virt.hardware [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 700.952782] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ce1c3f-a76f-4cfa-9398-f3850a162a5f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.965023] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf49aa5b-ff59-40b7-be3a-bb693c602176 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.981904] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.094s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.982422] env[61970]: ERROR nova.compute.manager [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9b7f5748-b3b2-41f9-ba64-c6efe0d7f78c, please check neutron logs for more information. [ 700.982422] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Traceback (most recent call last): [ 700.982422] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 700.982422] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] self.driver.spawn(context, instance, image_meta, [ 700.982422] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 700.982422] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] self._vmops.spawn(context, instance, image_meta, injected_files, [ 700.982422] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 700.982422] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] vm_ref = self.build_virtual_machine(instance, [ 700.982422] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 700.982422] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] vif_infos = vmwarevif.get_vif_info(self._session, [ 700.982422] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 700.982739] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] for vif in network_info: [ 700.982739] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 700.982739] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] return self._sync_wrapper(fn, *args, **kwargs) [ 700.982739] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 700.982739] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] self.wait() [ 700.982739] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 700.982739] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] self[:] = self._gt.wait() [ 700.982739] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 700.982739] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] return self._exit_event.wait() [ 700.982739] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 700.982739] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] current.throw(*self._exc) [ 700.982739] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 700.982739] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] result = function(*args, **kwargs) [ 700.983130] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 700.983130] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] return func(*args, **kwargs) [ 700.983130] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 700.983130] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] raise e [ 700.983130] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 700.983130] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] nwinfo = self.network_api.allocate_for_instance( [ 700.983130] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 700.983130] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] created_port_ids = self._update_ports_for_instance( [ 700.983130] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 700.983130] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] with excutils.save_and_reraise_exception(): [ 700.983130] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 700.983130] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] self.force_reraise() [ 700.983130] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 700.983450] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] raise self.value [ 700.983450] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 700.983450] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] updated_port = self._update_port( [ 700.983450] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 700.983450] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] _ensure_no_port_binding_failure(port) [ 700.983450] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 700.983450] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] raise exception.PortBindingFailed(port_id=port['id']) [ 700.983450] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] nova.exception.PortBindingFailed: Binding failed for port 9b7f5748-b3b2-41f9-ba64-c6efe0d7f78c, please check neutron logs for more information. [ 700.983450] env[61970]: ERROR nova.compute.manager [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] [ 700.983450] env[61970]: DEBUG nova.compute.utils [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Binding failed for port 9b7f5748-b3b2-41f9-ba64-c6efe0d7f78c, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 700.984361] env[61970]: DEBUG oslo_concurrency.lockutils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.695s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.985855] env[61970]: INFO nova.compute.claims [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 700.988854] env[61970]: DEBUG nova.compute.manager [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Build of instance 90cb79fb-aeb4-4c13-ab80-2787f67ff240 was re-scheduled: Binding failed for port 9b7f5748-b3b2-41f9-ba64-c6efe0d7f78c, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 700.989113] env[61970]: DEBUG nova.compute.manager [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 700.989258] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "refresh_cache-90cb79fb-aeb4-4c13-ab80-2787f67ff240" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.989420] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "refresh_cache-90cb79fb-aeb4-4c13-ab80-2787f67ff240" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.989901] env[61970]: DEBUG nova.network.neutron [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 701.015171] env[61970]: DEBUG nova.network.neutron [-] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.051575] env[61970]: INFO nova.compute.manager [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Took 40.27 seconds to build instance. [ 701.190933] env[61970]: ERROR nova.compute.manager [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ce17d99a-8184-48b2-a9ac-2afbf5f36def, please check neutron logs for more information. [ 701.190933] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 701.190933] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.190933] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 701.190933] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.190933] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 701.190933] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.190933] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 701.190933] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.190933] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 701.190933] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.190933] env[61970]: ERROR nova.compute.manager raise self.value [ 701.190933] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.190933] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 701.190933] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.190933] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 701.191537] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.191537] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 701.191537] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ce17d99a-8184-48b2-a9ac-2afbf5f36def, please check neutron logs for more information. [ 701.191537] env[61970]: ERROR nova.compute.manager [ 701.191537] env[61970]: Traceback (most recent call last): [ 701.191537] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 701.191537] env[61970]: listener.cb(fileno) [ 701.191537] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.191537] env[61970]: result = function(*args, **kwargs) [ 701.191537] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 701.191537] env[61970]: return func(*args, **kwargs) [ 701.191537] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.191537] env[61970]: raise e [ 701.191537] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.191537] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 701.191537] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.191537] env[61970]: created_port_ids = self._update_ports_for_instance( [ 701.191537] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.191537] env[61970]: with excutils.save_and_reraise_exception(): [ 701.191537] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.191537] env[61970]: self.force_reraise() [ 701.191537] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.191537] env[61970]: raise self.value [ 701.191537] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.191537] env[61970]: updated_port = self._update_port( [ 701.191537] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.191537] env[61970]: _ensure_no_port_binding_failure(port) [ 701.191537] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.191537] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 701.192520] env[61970]: nova.exception.PortBindingFailed: Binding failed for port ce17d99a-8184-48b2-a9ac-2afbf5f36def, please check neutron logs for more information. [ 701.192520] env[61970]: Removing descriptor: 17 [ 701.192520] env[61970]: ERROR nova.compute.manager [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ce17d99a-8184-48b2-a9ac-2afbf5f36def, please check neutron logs for more information. [ 701.192520] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Traceback (most recent call last): [ 701.192520] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 701.192520] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] yield resources [ 701.192520] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 701.192520] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] self.driver.spawn(context, instance, image_meta, [ 701.192520] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 701.192520] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.192520] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.192520] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] vm_ref = self.build_virtual_machine(instance, [ 701.192930] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.192930] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.192930] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.192930] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] for vif in network_info: [ 701.192930] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 701.192930] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] return self._sync_wrapper(fn, *args, **kwargs) [ 701.192930] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 701.192930] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] self.wait() [ 701.192930] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 701.192930] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] self[:] = self._gt.wait() [ 701.192930] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.192930] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] return self._exit_event.wait() [ 701.192930] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 701.193402] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] result = hub.switch() [ 701.193402] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 701.193402] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] return self.greenlet.switch() [ 701.193402] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.193402] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] result = function(*args, **kwargs) [ 701.193402] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 701.193402] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] return func(*args, **kwargs) [ 701.193402] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.193402] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] raise e [ 701.193402] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.193402] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] nwinfo = self.network_api.allocate_for_instance( [ 701.193402] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.193402] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] created_port_ids = self._update_ports_for_instance( [ 701.193925] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.193925] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] with excutils.save_and_reraise_exception(): [ 701.193925] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.193925] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] self.force_reraise() [ 701.193925] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.193925] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] raise self.value [ 701.193925] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.193925] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] updated_port = self._update_port( [ 701.193925] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.193925] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] _ensure_no_port_binding_failure(port) [ 701.193925] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.193925] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] raise exception.PortBindingFailed(port_id=port['id']) [ 701.194386] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] nova.exception.PortBindingFailed: Binding failed for port ce17d99a-8184-48b2-a9ac-2afbf5f36def, please check neutron logs for more information. [ 701.194386] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] [ 701.194386] env[61970]: INFO nova.compute.manager [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Terminating instance [ 701.194386] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Acquiring lock "refresh_cache-7dda08a4-83ae-4882-9469-324d948c9749" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.194818] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Acquired lock "refresh_cache-7dda08a4-83ae-4882-9469-324d948c9749" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.194947] env[61970]: DEBUG nova.network.neutron [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 701.508778] env[61970]: DEBUG nova.network.neutron [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.520113] env[61970]: INFO nova.compute.manager [-] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Took 1.03 seconds to deallocate network for instance. [ 701.522473] env[61970]: DEBUG nova.compute.claims [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 701.522651] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.553266] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44039091-1c0a-431c-93ce-53aac0f9a6c7 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Lock "9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.279s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.590528] env[61970]: DEBUG nova.network.neutron [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.712827] env[61970]: DEBUG nova.network.neutron [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.793577] env[61970]: DEBUG nova.network.neutron [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.930068] env[61970]: INFO nova.compute.manager [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Rebuilding instance [ 701.972673] env[61970]: DEBUG nova.compute.manager [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 701.973553] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ba07ca-6c05-46e7-a42d-9416da653f70 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.055637] env[61970]: DEBUG nova.compute.manager [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 702.092684] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "refresh_cache-90cb79fb-aeb4-4c13-ab80-2787f67ff240" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.093033] env[61970]: DEBUG nova.compute.manager [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 702.093103] env[61970]: DEBUG nova.compute.manager [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 702.093682] env[61970]: DEBUG nova.network.neutron [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 702.130743] env[61970]: DEBUG nova.network.neutron [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.173144] env[61970]: DEBUG nova.compute.manager [req-40b4a646-3405-45fc-ae33-d0212b09a7ae req-10819d2f-eb89-4e5a-9669-724547d31460 service nova] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Received event network-changed-ce17d99a-8184-48b2-a9ac-2afbf5f36def {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 702.173347] env[61970]: DEBUG nova.compute.manager [req-40b4a646-3405-45fc-ae33-d0212b09a7ae req-10819d2f-eb89-4e5a-9669-724547d31460 service nova] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Refreshing instance network info cache due to event network-changed-ce17d99a-8184-48b2-a9ac-2afbf5f36def. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 702.173585] env[61970]: DEBUG oslo_concurrency.lockutils [req-40b4a646-3405-45fc-ae33-d0212b09a7ae req-10819d2f-eb89-4e5a-9669-724547d31460 service nova] Acquiring lock "refresh_cache-7dda08a4-83ae-4882-9469-324d948c9749" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.296528] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Releasing lock "refresh_cache-7dda08a4-83ae-4882-9469-324d948c9749" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.296947] env[61970]: DEBUG nova.compute.manager [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 702.297158] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 702.297468] env[61970]: DEBUG oslo_concurrency.lockutils [req-40b4a646-3405-45fc-ae33-d0212b09a7ae req-10819d2f-eb89-4e5a-9669-724547d31460 service nova] Acquired lock "refresh_cache-7dda08a4-83ae-4882-9469-324d948c9749" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.297635] env[61970]: DEBUG nova.network.neutron [req-40b4a646-3405-45fc-ae33-d0212b09a7ae req-10819d2f-eb89-4e5a-9669-724547d31460 service nova] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Refreshing network info cache for port ce17d99a-8184-48b2-a9ac-2afbf5f36def {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 702.298708] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ee718a59-a602-45e3-bd8b-feebff88b3e8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.309166] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fac015e-ce1d-4adf-90f4-683d4306cb20 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.335200] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7dda08a4-83ae-4882-9469-324d948c9749 could not be found. [ 702.335433] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 702.335610] env[61970]: INFO nova.compute.manager [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Took 0.04 seconds to destroy the instance on the hypervisor. [ 702.335852] env[61970]: DEBUG oslo.service.loopingcall [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 702.339035] env[61970]: DEBUG nova.compute.manager [-] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 702.339035] env[61970]: DEBUG nova.network.neutron [-] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 702.356546] env[61970]: DEBUG nova.network.neutron [-] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.436359] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-873af809-1738-4bfa-a1d3-2cf392ca94e3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.443952] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d95207a0-3417-4f3e-ae51-4fc152707657 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.474541] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae02ff4-fefa-40e8-a667-8b1c295d8b5f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.481890] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fdd0f5c-d070-4e3c-9a9d-f19a575a399e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.487011] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 702.487280] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0cd5609b-df4b-4539-9b9b-bddda0c55bea {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.499023] env[61970]: DEBUG nova.compute.provider_tree [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.501826] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 702.501826] env[61970]: value = "task-1355507" [ 702.501826] env[61970]: _type = "Task" [ 702.501826] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.510273] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355507, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.573510] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.635794] env[61970]: DEBUG nova.network.neutron [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.820605] env[61970]: DEBUG nova.network.neutron [req-40b4a646-3405-45fc-ae33-d0212b09a7ae req-10819d2f-eb89-4e5a-9669-724547d31460 service nova] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.860258] env[61970]: DEBUG nova.network.neutron [-] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.919972] env[61970]: DEBUG nova.network.neutron [req-40b4a646-3405-45fc-ae33-d0212b09a7ae req-10819d2f-eb89-4e5a-9669-724547d31460 service nova] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.003917] env[61970]: DEBUG nova.scheduler.client.report [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.016619] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355507, 'name': PowerOffVM_Task, 'duration_secs': 0.171723} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.016902] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 703.017139] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 703.017951] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4de9064-ad2a-4ad9-8462-16aa98d4a11c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.025021] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 703.025120] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1fbe270-4172-4b3d-a69d-a5e0b4643e7e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.050516] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 703.050516] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 703.050516] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Deleting the datastore file [datastore2] 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 703.050786] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bacc510c-b8c7-47f7-8c35-40cba8a16ebe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.057635] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 703.057635] env[61970]: value = "task-1355509" [ 703.057635] env[61970]: _type = "Task" [ 703.057635] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.065561] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355509, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.139345] env[61970]: INFO nova.compute.manager [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 90cb79fb-aeb4-4c13-ab80-2787f67ff240] Took 1.05 seconds to deallocate network for instance. [ 703.363118] env[61970]: INFO nova.compute.manager [-] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Took 1.02 seconds to deallocate network for instance. [ 703.365524] env[61970]: DEBUG nova.compute.claims [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 703.365701] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.422503] env[61970]: DEBUG oslo_concurrency.lockutils [req-40b4a646-3405-45fc-ae33-d0212b09a7ae req-10819d2f-eb89-4e5a-9669-724547d31460 service nova] Releasing lock "refresh_cache-7dda08a4-83ae-4882-9469-324d948c9749" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.422786] env[61970]: DEBUG nova.compute.manager [req-40b4a646-3405-45fc-ae33-d0212b09a7ae req-10819d2f-eb89-4e5a-9669-724547d31460 service nova] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Received event network-vif-deleted-ce17d99a-8184-48b2-a9ac-2afbf5f36def {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 703.513802] env[61970]: DEBUG oslo_concurrency.lockutils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.528s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.514147] env[61970]: DEBUG oslo_concurrency.lockutils [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.201s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.568222] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355509, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.086637} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.568577] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 703.568881] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 703.569113] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 704.019717] env[61970]: DEBUG oslo_concurrency.lockutils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Acquiring lock "5e7d7c83-9365-462e-a417-797bde525a54" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.020024] env[61970]: DEBUG oslo_concurrency.lockutils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Lock "5e7d7c83-9365-462e-a417-797bde525a54" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.166774] env[61970]: INFO nova.scheduler.client.report [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Deleted allocations for instance 90cb79fb-aeb4-4c13-ab80-2787f67ff240 [ 704.413268] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c40843-051a-4839-8cb5-10101dbebae4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.422292] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e1aa782-aa9f-4713-859b-d4b4c23c8daa {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.451069] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe4ccf9-49aa-4f58-b108-0f338cafb76c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.458846] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc14484-a6ac-43cc-bd0d-73ad094e70ec {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.472149] env[61970]: DEBUG nova.compute.provider_tree [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.524977] env[61970]: DEBUG oslo_concurrency.lockutils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Lock "5e7d7c83-9365-462e-a417-797bde525a54" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.505s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.525704] env[61970]: DEBUG nova.compute.manager [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 704.603104] env[61970]: DEBUG nova.virt.hardware [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 704.603344] env[61970]: DEBUG nova.virt.hardware [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 704.603500] env[61970]: DEBUG nova.virt.hardware [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 704.603681] env[61970]: DEBUG nova.virt.hardware [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 704.603834] env[61970]: DEBUG nova.virt.hardware [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 704.603969] env[61970]: DEBUG nova.virt.hardware [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 704.604188] env[61970]: DEBUG nova.virt.hardware [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 704.604346] env[61970]: DEBUG nova.virt.hardware [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 704.604508] env[61970]: DEBUG nova.virt.hardware [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 704.604666] env[61970]: DEBUG nova.virt.hardware [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 704.605247] env[61970]: DEBUG nova.virt.hardware [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 704.605692] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f038713d-7fab-4f27-8705-543b55b341ae {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.615191] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f915b5d-61f7-42ed-ad0e-abb64b8bfcbe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.628409] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Instance VIF info [] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 704.633983] env[61970]: DEBUG oslo.service.loopingcall [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 704.634120] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 704.635031] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-82a27b4e-4305-4e18-81cb-cdb3bc82b551 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.651109] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 704.651109] env[61970]: value = "task-1355510" [ 704.651109] env[61970]: _type = "Task" [ 704.651109] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.658590] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355510, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.675292] env[61970]: DEBUG oslo_concurrency.lockutils [None req-69efc960-8e85-4279-881f-521601f07e6b tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "90cb79fb-aeb4-4c13-ab80-2787f67ff240" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.701s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.975614] env[61970]: DEBUG nova.scheduler.client.report [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.030990] env[61970]: DEBUG nova.compute.utils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 705.032423] env[61970]: DEBUG nova.compute.manager [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 705.032597] env[61970]: DEBUG nova.network.neutron [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 705.073056] env[61970]: DEBUG nova.policy [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '23b5d2d68f784317a200e5b8217fb6bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0737431206af4f0780db980c7b27368e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 705.162121] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355510, 'name': CreateVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.178406] env[61970]: DEBUG nova.compute.manager [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 705.391678] env[61970]: DEBUG nova.network.neutron [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Successfully created port: da878503-b70a-4e57-b4d1-eacf4bb3c3f3 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 705.480811] env[61970]: DEBUG oslo_concurrency.lockutils [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.967s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.481463] env[61970]: ERROR nova.compute.manager [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d00b221d-7660-4f28-a08a-e39ff33fd1ee, please check neutron logs for more information. [ 705.481463] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Traceback (most recent call last): [ 705.481463] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 705.481463] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] self.driver.spawn(context, instance, image_meta, [ 705.481463] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 705.481463] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 705.481463] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 705.481463] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] vm_ref = self.build_virtual_machine(instance, [ 705.481463] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 705.481463] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] vif_infos = vmwarevif.get_vif_info(self._session, [ 705.481463] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 705.481827] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] for vif in network_info: [ 705.481827] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 705.481827] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] return self._sync_wrapper(fn, *args, **kwargs) [ 705.481827] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 705.481827] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] self.wait() [ 705.481827] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 705.481827] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] self[:] = self._gt.wait() [ 705.481827] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 705.481827] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] return self._exit_event.wait() [ 705.481827] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 705.481827] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] current.throw(*self._exc) [ 705.481827] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.481827] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] result = function(*args, **kwargs) [ 705.482215] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 705.482215] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] return func(*args, **kwargs) [ 705.482215] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.482215] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] raise e [ 705.482215] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.482215] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] nwinfo = self.network_api.allocate_for_instance( [ 705.482215] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.482215] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] created_port_ids = self._update_ports_for_instance( [ 705.482215] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.482215] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] with excutils.save_and_reraise_exception(): [ 705.482215] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.482215] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] self.force_reraise() [ 705.482215] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.482563] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] raise self.value [ 705.482563] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.482563] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] updated_port = self._update_port( [ 705.482563] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.482563] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] _ensure_no_port_binding_failure(port) [ 705.482563] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.482563] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] raise exception.PortBindingFailed(port_id=port['id']) [ 705.482563] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] nova.exception.PortBindingFailed: Binding failed for port d00b221d-7660-4f28-a08a-e39ff33fd1ee, please check neutron logs for more information. [ 705.482563] env[61970]: ERROR nova.compute.manager [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] [ 705.482563] env[61970]: DEBUG nova.compute.utils [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Binding failed for port d00b221d-7660-4f28-a08a-e39ff33fd1ee, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 705.488697] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.640s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.489490] env[61970]: INFO nova.compute.claims [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 705.496429] env[61970]: DEBUG nova.compute.manager [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Build of instance 719de3af-ffe0-4352-9d2c-9e102859fb6c was re-scheduled: Binding failed for port d00b221d-7660-4f28-a08a-e39ff33fd1ee, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 705.500020] env[61970]: DEBUG nova.compute.manager [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 705.500020] env[61970]: DEBUG oslo_concurrency.lockutils [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "refresh_cache-719de3af-ffe0-4352-9d2c-9e102859fb6c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.500020] env[61970]: DEBUG oslo_concurrency.lockutils [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "refresh_cache-719de3af-ffe0-4352-9d2c-9e102859fb6c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.500020] env[61970]: DEBUG nova.network.neutron [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 705.535716] env[61970]: DEBUG nova.compute.manager [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 705.662441] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355510, 'name': CreateVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.713807] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.023349] env[61970]: DEBUG nova.network.neutron [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.146118] env[61970]: DEBUG nova.network.neutron [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.165753] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355510, 'name': CreateVM_Task, 'duration_secs': 1.254453} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.166077] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 706.166426] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.166511] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.166770] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 706.167018] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb96f906-71aa-4470-8ccb-e02808c0c511 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.171731] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 706.171731] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523f552a-b741-a687-b5bc-de00e10faa31" [ 706.171731] env[61970]: _type = "Task" [ 706.171731] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.179960] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523f552a-b741-a687-b5bc-de00e10faa31, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.315181] env[61970]: DEBUG nova.compute.manager [req-b6da20c0-abad-4b66-806c-83b1d686195d req-a297bf65-baa8-4ce4-9fa8-fc7a1687d439 service nova] [instance: 9d71be92-36ef-4523-919a-931110d73248] Received event network-changed-da878503-b70a-4e57-b4d1-eacf4bb3c3f3 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 706.315181] env[61970]: DEBUG nova.compute.manager [req-b6da20c0-abad-4b66-806c-83b1d686195d req-a297bf65-baa8-4ce4-9fa8-fc7a1687d439 service nova] [instance: 9d71be92-36ef-4523-919a-931110d73248] Refreshing instance network info cache due to event network-changed-da878503-b70a-4e57-b4d1-eacf4bb3c3f3. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 706.315181] env[61970]: DEBUG oslo_concurrency.lockutils [req-b6da20c0-abad-4b66-806c-83b1d686195d req-a297bf65-baa8-4ce4-9fa8-fc7a1687d439 service nova] Acquiring lock "refresh_cache-9d71be92-36ef-4523-919a-931110d73248" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.315181] env[61970]: DEBUG oslo_concurrency.lockutils [req-b6da20c0-abad-4b66-806c-83b1d686195d req-a297bf65-baa8-4ce4-9fa8-fc7a1687d439 service nova] Acquired lock "refresh_cache-9d71be92-36ef-4523-919a-931110d73248" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.315181] env[61970]: DEBUG nova.network.neutron [req-b6da20c0-abad-4b66-806c-83b1d686195d req-a297bf65-baa8-4ce4-9fa8-fc7a1687d439 service nova] [instance: 9d71be92-36ef-4523-919a-931110d73248] Refreshing network info cache for port da878503-b70a-4e57-b4d1-eacf4bb3c3f3 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 706.451010] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "b19cb516-b163-4bed-ba5b-139a0a18fc05" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.451265] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "b19cb516-b163-4bed-ba5b-139a0a18fc05" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.516098] env[61970]: ERROR nova.compute.manager [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port da878503-b70a-4e57-b4d1-eacf4bb3c3f3, please check neutron logs for more information. [ 706.516098] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 706.516098] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.516098] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 706.516098] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.516098] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 706.516098] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.516098] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 706.516098] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.516098] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 706.516098] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.516098] env[61970]: ERROR nova.compute.manager raise self.value [ 706.516098] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.516098] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 706.516098] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.516098] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 706.516622] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.516622] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 706.516622] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port da878503-b70a-4e57-b4d1-eacf4bb3c3f3, please check neutron logs for more information. [ 706.516622] env[61970]: ERROR nova.compute.manager [ 706.516622] env[61970]: Traceback (most recent call last): [ 706.516622] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 706.516622] env[61970]: listener.cb(fileno) [ 706.516622] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.516622] env[61970]: result = function(*args, **kwargs) [ 706.516622] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 706.516622] env[61970]: return func(*args, **kwargs) [ 706.516622] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.516622] env[61970]: raise e [ 706.516622] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.516622] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 706.516622] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.516622] env[61970]: created_port_ids = self._update_ports_for_instance( [ 706.516622] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.516622] env[61970]: with excutils.save_and_reraise_exception(): [ 706.516622] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.516622] env[61970]: self.force_reraise() [ 706.516622] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.516622] env[61970]: raise self.value [ 706.516622] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.516622] env[61970]: updated_port = self._update_port( [ 706.516622] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.516622] env[61970]: _ensure_no_port_binding_failure(port) [ 706.516622] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.516622] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 706.517450] env[61970]: nova.exception.PortBindingFailed: Binding failed for port da878503-b70a-4e57-b4d1-eacf4bb3c3f3, please check neutron logs for more information. [ 706.517450] env[61970]: Removing descriptor: 17 [ 706.546248] env[61970]: DEBUG nova.compute.manager [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 706.575088] env[61970]: DEBUG nova.virt.hardware [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 706.575337] env[61970]: DEBUG nova.virt.hardware [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 706.575491] env[61970]: DEBUG nova.virt.hardware [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 706.575671] env[61970]: DEBUG nova.virt.hardware [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 706.575814] env[61970]: DEBUG nova.virt.hardware [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 706.575961] env[61970]: DEBUG nova.virt.hardware [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 706.576183] env[61970]: DEBUG nova.virt.hardware [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 706.576342] env[61970]: DEBUG nova.virt.hardware [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 706.576504] env[61970]: DEBUG nova.virt.hardware [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 706.576675] env[61970]: DEBUG nova.virt.hardware [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 706.576891] env[61970]: DEBUG nova.virt.hardware [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 706.577799] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dcd0a94-9060-4e51-82c4-7d68359f4889 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.587062] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f7fad0-6800-45fb-8067-de3b37fe8480 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.602667] env[61970]: ERROR nova.compute.manager [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port da878503-b70a-4e57-b4d1-eacf4bb3c3f3, please check neutron logs for more information. [ 706.602667] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] Traceback (most recent call last): [ 706.602667] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 706.602667] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] yield resources [ 706.602667] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 706.602667] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] self.driver.spawn(context, instance, image_meta, [ 706.602667] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 706.602667] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] self._vmops.spawn(context, instance, image_meta, injected_files, [ 706.602667] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 706.602667] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] vm_ref = self.build_virtual_machine(instance, [ 706.602667] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 706.603053] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] vif_infos = vmwarevif.get_vif_info(self._session, [ 706.603053] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 706.603053] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] for vif in network_info: [ 706.603053] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 706.603053] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] return self._sync_wrapper(fn, *args, **kwargs) [ 706.603053] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 706.603053] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] self.wait() [ 706.603053] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 706.603053] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] self[:] = self._gt.wait() [ 706.603053] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 706.603053] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] return self._exit_event.wait() [ 706.603053] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 706.603053] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] current.throw(*self._exc) [ 706.603424] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.603424] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] result = function(*args, **kwargs) [ 706.603424] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 706.603424] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] return func(*args, **kwargs) [ 706.603424] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.603424] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] raise e [ 706.603424] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.603424] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] nwinfo = self.network_api.allocate_for_instance( [ 706.603424] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.603424] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] created_port_ids = self._update_ports_for_instance( [ 706.603424] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.603424] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] with excutils.save_and_reraise_exception(): [ 706.603424] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.603798] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] self.force_reraise() [ 706.603798] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.603798] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] raise self.value [ 706.603798] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.603798] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] updated_port = self._update_port( [ 706.603798] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.603798] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] _ensure_no_port_binding_failure(port) [ 706.603798] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.603798] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] raise exception.PortBindingFailed(port_id=port['id']) [ 706.603798] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] nova.exception.PortBindingFailed: Binding failed for port da878503-b70a-4e57-b4d1-eacf4bb3c3f3, please check neutron logs for more information. [ 706.603798] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] [ 706.603798] env[61970]: INFO nova.compute.manager [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Terminating instance [ 706.606210] env[61970]: DEBUG oslo_concurrency.lockutils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Acquiring lock "refresh_cache-9d71be92-36ef-4523-919a-931110d73248" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.650392] env[61970]: DEBUG oslo_concurrency.lockutils [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "refresh_cache-719de3af-ffe0-4352-9d2c-9e102859fb6c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.650617] env[61970]: DEBUG nova.compute.manager [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 706.650790] env[61970]: DEBUG nova.compute.manager [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 706.650954] env[61970]: DEBUG nova.network.neutron [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 706.668862] env[61970]: DEBUG nova.network.neutron [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.682137] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523f552a-b741-a687-b5bc-de00e10faa31, 'name': SearchDatastore_Task, 'duration_secs': 0.010762} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.685125] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.685365] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 706.685585] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.685728] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.685903] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 706.686571] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d6ef75c9-0cc4-40cb-8668-f1bd5a429906 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.694288] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 706.694421] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 706.695148] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24e670e0-4adb-421f-8eea-a0973215aac9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.702913] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 706.702913] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d39050-85e3-7e78-cf78-966c89f5ce74" [ 706.702913] env[61970]: _type = "Task" [ 706.702913] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.710329] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d39050-85e3-7e78-cf78-966c89f5ce74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.831637] env[61970]: DEBUG nova.network.neutron [req-b6da20c0-abad-4b66-806c-83b1d686195d req-a297bf65-baa8-4ce4-9fa8-fc7a1687d439 service nova] [instance: 9d71be92-36ef-4523-919a-931110d73248] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.921834] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-074f2a3d-e19f-4288-9476-3d001ee9f9cb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.924044] env[61970]: DEBUG nova.network.neutron [req-b6da20c0-abad-4b66-806c-83b1d686195d req-a297bf65-baa8-4ce4-9fa8-fc7a1687d439 service nova] [instance: 9d71be92-36ef-4523-919a-931110d73248] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.932109] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6a030b-8be6-442d-aa35-f82924361335 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.965642] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e2f6d0-d650-4480-a7a5-6a9fdfdfc60d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.973229] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3e44c3-4385-40e2-95e9-95ac4f3ac001 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.987373] env[61970]: DEBUG nova.compute.provider_tree [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.172602] env[61970]: DEBUG nova.network.neutron [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.212526] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d39050-85e3-7e78-cf78-966c89f5ce74, 'name': SearchDatastore_Task, 'duration_secs': 0.007739} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.213303] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a61f24c-d23b-4d22-a26b-f317566031af {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.218789] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 707.218789] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529e04c0-7208-ed01-f277-d5be3e224e7f" [ 707.218789] env[61970]: _type = "Task" [ 707.218789] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.225324] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529e04c0-7208-ed01-f277-d5be3e224e7f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.427239] env[61970]: DEBUG oslo_concurrency.lockutils [req-b6da20c0-abad-4b66-806c-83b1d686195d req-a297bf65-baa8-4ce4-9fa8-fc7a1687d439 service nova] Releasing lock "refresh_cache-9d71be92-36ef-4523-919a-931110d73248" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.427705] env[61970]: DEBUG oslo_concurrency.lockutils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Acquired lock "refresh_cache-9d71be92-36ef-4523-919a-931110d73248" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.427928] env[61970]: DEBUG nova.network.neutron [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.490787] env[61970]: DEBUG nova.scheduler.client.report [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.675115] env[61970]: INFO nova.compute.manager [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: 719de3af-ffe0-4352-9d2c-9e102859fb6c] Took 1.02 seconds to deallocate network for instance. [ 707.729767] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529e04c0-7208-ed01-f277-d5be3e224e7f, 'name': SearchDatastore_Task, 'duration_secs': 0.008352} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.730126] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.730385] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8/9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 707.730641] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9f6fc96-f702-4751-b1bf-3ab2e53364d2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.738197] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 707.738197] env[61970]: value = "task-1355511" [ 707.738197] env[61970]: _type = "Task" [ 707.738197] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.745727] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355511, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.953603] env[61970]: DEBUG nova.network.neutron [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.995447] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.507s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.995787] env[61970]: DEBUG nova.compute.manager [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 708.000322] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.684s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.052381] env[61970]: DEBUG nova.network.neutron [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.248401] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355511, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.47854} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.248662] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8/9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 708.248933] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 708.249217] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb1b6270-a5cd-4b9e-b7f4-2458abe97b7a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.256504] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 708.256504] env[61970]: value = "task-1355512" [ 708.256504] env[61970]: _type = "Task" [ 708.256504] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.264265] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355512, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.353727] env[61970]: DEBUG nova.compute.manager [req-0ac189ba-cb55-4d22-af28-30b9b60b2bda req-9ffa882f-f8b3-40ef-bdd9-c6c5218f1e71 service nova] [instance: 9d71be92-36ef-4523-919a-931110d73248] Received event network-vif-deleted-da878503-b70a-4e57-b4d1-eacf4bb3c3f3 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 708.505190] env[61970]: DEBUG nova.compute.utils [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 708.506554] env[61970]: DEBUG nova.compute.manager [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 708.506814] env[61970]: DEBUG nova.network.neutron [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 708.555102] env[61970]: DEBUG oslo_concurrency.lockutils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Releasing lock "refresh_cache-9d71be92-36ef-4523-919a-931110d73248" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.555545] env[61970]: DEBUG nova.compute.manager [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 708.555799] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 708.556125] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-de205986-22d1-4a73-9ce6-c50d05e39bce {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.563033] env[61970]: DEBUG nova.policy [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0416ea0a7c864ac3a4fa38987c6efde9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ce5bd66f3a24856aa30d234659dfc1d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 708.568387] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cb87084-845d-403f-ac41-1599cda3561f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.593477] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9d71be92-36ef-4523-919a-931110d73248 could not be found. [ 708.593702] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 708.593878] env[61970]: INFO nova.compute.manager [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Took 0.04 seconds to destroy the instance on the hypervisor. [ 708.594142] env[61970]: DEBUG oslo.service.loopingcall [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.596771] env[61970]: DEBUG nova.compute.manager [-] [instance: 9d71be92-36ef-4523-919a-931110d73248] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.596843] env[61970]: DEBUG nova.network.neutron [-] [instance: 9d71be92-36ef-4523-919a-931110d73248] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 708.617012] env[61970]: DEBUG nova.network.neutron [-] [instance: 9d71be92-36ef-4523-919a-931110d73248] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.713987] env[61970]: INFO nova.scheduler.client.report [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Deleted allocations for instance 719de3af-ffe0-4352-9d2c-9e102859fb6c [ 708.766816] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355512, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076865} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.769350] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 708.771381] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58980c2f-7921-4229-ae87-c46a82fc29ed {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.796354] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Reconfiguring VM instance instance-00000022 to attach disk [datastore2] 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8/9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 708.796957] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c02a536c-1946-4a38-8d38-7c18278fac1e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.826017] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 708.826017] env[61970]: value = "task-1355513" [ 708.826017] env[61970]: _type = "Task" [ 708.826017] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.831649] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355513, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.926441] env[61970]: DEBUG nova.network.neutron [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Successfully created port: b758644c-56b6-4a94-b6e6-0d038fd951b4 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 709.012845] env[61970]: DEBUG nova.compute.manager [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 709.015788] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c334793d-926d-4285-92b1-6541e26eaf42 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.024832] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956c68da-9aa1-4f39-b9af-17dafe15f21b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.056545] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9c6389-2649-4504-b4e7-e074a97d3cf0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.065178] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e12af5-391a-4432-bc8f-62da39443b5d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.082449] env[61970]: DEBUG nova.compute.provider_tree [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.120684] env[61970]: DEBUG nova.network.neutron [-] [instance: 9d71be92-36ef-4523-919a-931110d73248] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.228928] env[61970]: DEBUG oslo_concurrency.lockutils [None req-408451dd-0ef2-46a6-91b9-f65be47487b6 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "719de3af-ffe0-4352-9d2c-9e102859fb6c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.369s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.334784] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355513, 'name': ReconfigVM_Task, 'duration_secs': 0.269728} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.335073] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Reconfigured VM instance instance-00000022 to attach disk [datastore2] 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8/9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 709.335828] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f93db1bb-0b80-4330-b013-938c221ce784 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.341799] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 709.341799] env[61970]: value = "task-1355514" [ 709.341799] env[61970]: _type = "Task" [ 709.341799] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.353691] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355514, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.585493] env[61970]: DEBUG nova.scheduler.client.report [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 709.623994] env[61970]: INFO nova.compute.manager [-] [instance: 9d71be92-36ef-4523-919a-931110d73248] Took 1.03 seconds to deallocate network for instance. [ 709.627534] env[61970]: DEBUG nova.compute.claims [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 709.627534] env[61970]: DEBUG oslo_concurrency.lockutils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.730959] env[61970]: DEBUG nova.compute.manager [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 709.851595] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355514, 'name': Rename_Task, 'duration_secs': 0.128737} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.851890] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 709.852134] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab44d120-0354-4cb5-beeb-8c573b795fa6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.863382] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 709.863382] env[61970]: value = "task-1355515" [ 709.863382] env[61970]: _type = "Task" [ 709.863382] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.871167] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355515, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.949017] env[61970]: ERROR nova.compute.manager [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b758644c-56b6-4a94-b6e6-0d038fd951b4, please check neutron logs for more information. [ 709.949017] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 709.949017] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 709.949017] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 709.949017] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.949017] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 709.949017] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.949017] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 709.949017] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.949017] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 709.949017] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.949017] env[61970]: ERROR nova.compute.manager raise self.value [ 709.949017] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.949017] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 709.949017] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.949017] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 709.949504] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.949504] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 709.949504] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b758644c-56b6-4a94-b6e6-0d038fd951b4, please check neutron logs for more information. [ 709.949504] env[61970]: ERROR nova.compute.manager [ 709.949504] env[61970]: Traceback (most recent call last): [ 709.949504] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 709.949504] env[61970]: listener.cb(fileno) [ 709.949504] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 709.949504] env[61970]: result = function(*args, **kwargs) [ 709.949504] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 709.949504] env[61970]: return func(*args, **kwargs) [ 709.949504] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 709.949504] env[61970]: raise e [ 709.949504] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 709.949504] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 709.949504] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.949504] env[61970]: created_port_ids = self._update_ports_for_instance( [ 709.949504] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.949504] env[61970]: with excutils.save_and_reraise_exception(): [ 709.949504] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.949504] env[61970]: self.force_reraise() [ 709.949504] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.949504] env[61970]: raise self.value [ 709.949504] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.949504] env[61970]: updated_port = self._update_port( [ 709.949504] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.949504] env[61970]: _ensure_no_port_binding_failure(port) [ 709.949504] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.949504] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 709.950339] env[61970]: nova.exception.PortBindingFailed: Binding failed for port b758644c-56b6-4a94-b6e6-0d038fd951b4, please check neutron logs for more information. [ 709.950339] env[61970]: Removing descriptor: 17 [ 710.026761] env[61970]: DEBUG nova.compute.manager [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 710.053773] env[61970]: DEBUG nova.virt.hardware [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 710.054045] env[61970]: DEBUG nova.virt.hardware [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 710.054186] env[61970]: DEBUG nova.virt.hardware [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 710.054367] env[61970]: DEBUG nova.virt.hardware [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 710.054511] env[61970]: DEBUG nova.virt.hardware [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 710.054656] env[61970]: DEBUG nova.virt.hardware [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 710.054860] env[61970]: DEBUG nova.virt.hardware [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 710.055031] env[61970]: DEBUG nova.virt.hardware [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 710.055201] env[61970]: DEBUG nova.virt.hardware [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 710.055361] env[61970]: DEBUG nova.virt.hardware [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 710.055530] env[61970]: DEBUG nova.virt.hardware [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 710.056408] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14374773-ad2d-461b-bc5c-e4583b005d7c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.064658] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-056ee080-cd25-4c4c-ad1b-60bd355bca54 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.078278] env[61970]: ERROR nova.compute.manager [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b758644c-56b6-4a94-b6e6-0d038fd951b4, please check neutron logs for more information. [ 710.078278] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Traceback (most recent call last): [ 710.078278] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 710.078278] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] yield resources [ 710.078278] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 710.078278] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] self.driver.spawn(context, instance, image_meta, [ 710.078278] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 710.078278] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 710.078278] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 710.078278] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] vm_ref = self.build_virtual_machine(instance, [ 710.078278] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 710.078708] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] vif_infos = vmwarevif.get_vif_info(self._session, [ 710.078708] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 710.078708] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] for vif in network_info: [ 710.078708] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 710.078708] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] return self._sync_wrapper(fn, *args, **kwargs) [ 710.078708] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 710.078708] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] self.wait() [ 710.078708] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 710.078708] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] self[:] = self._gt.wait() [ 710.078708] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 710.078708] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] return self._exit_event.wait() [ 710.078708] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 710.078708] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] current.throw(*self._exc) [ 710.079157] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.079157] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] result = function(*args, **kwargs) [ 710.079157] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 710.079157] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] return func(*args, **kwargs) [ 710.079157] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.079157] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] raise e [ 710.079157] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.079157] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] nwinfo = self.network_api.allocate_for_instance( [ 710.079157] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.079157] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] created_port_ids = self._update_ports_for_instance( [ 710.079157] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.079157] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] with excutils.save_and_reraise_exception(): [ 710.079157] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.079561] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] self.force_reraise() [ 710.079561] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.079561] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] raise self.value [ 710.079561] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.079561] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] updated_port = self._update_port( [ 710.079561] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.079561] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] _ensure_no_port_binding_failure(port) [ 710.079561] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.079561] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] raise exception.PortBindingFailed(port_id=port['id']) [ 710.079561] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] nova.exception.PortBindingFailed: Binding failed for port b758644c-56b6-4a94-b6e6-0d038fd951b4, please check neutron logs for more information. [ 710.079561] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] [ 710.079561] env[61970]: INFO nova.compute.manager [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Terminating instance [ 710.080717] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Acquiring lock "refresh_cache-630145f3-cd01-41fe-b9c5-7ebf44b1e6d9" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.080871] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Acquired lock "refresh_cache-630145f3-cd01-41fe-b9c5-7ebf44b1e6d9" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.081052] env[61970]: DEBUG nova.network.neutron [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 710.090326] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.090s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.090964] env[61970]: ERROR nova.compute.manager [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port df4144c7-1e44-42cf-a3fe-289f5783650d, please check neutron logs for more information. [ 710.090964] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Traceback (most recent call last): [ 710.090964] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 710.090964] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] self.driver.spawn(context, instance, image_meta, [ 710.090964] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 710.090964] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 710.090964] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 710.090964] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] vm_ref = self.build_virtual_machine(instance, [ 710.090964] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 710.090964] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] vif_infos = vmwarevif.get_vif_info(self._session, [ 710.090964] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 710.091364] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] for vif in network_info: [ 710.091364] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 710.091364] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] return self._sync_wrapper(fn, *args, **kwargs) [ 710.091364] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 710.091364] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] self.wait() [ 710.091364] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 710.091364] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] self[:] = self._gt.wait() [ 710.091364] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 710.091364] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] return self._exit_event.wait() [ 710.091364] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 710.091364] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] result = hub.switch() [ 710.091364] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 710.091364] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] return self.greenlet.switch() [ 710.091790] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.091790] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] result = function(*args, **kwargs) [ 710.091790] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 710.091790] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] return func(*args, **kwargs) [ 710.091790] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.091790] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] raise e [ 710.091790] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.091790] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] nwinfo = self.network_api.allocate_for_instance( [ 710.091790] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.091790] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] created_port_ids = self._update_ports_for_instance( [ 710.091790] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.091790] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] with excutils.save_and_reraise_exception(): [ 710.091790] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.092178] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] self.force_reraise() [ 710.092178] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.092178] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] raise self.value [ 710.092178] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.092178] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] updated_port = self._update_port( [ 710.092178] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.092178] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] _ensure_no_port_binding_failure(port) [ 710.092178] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.092178] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] raise exception.PortBindingFailed(port_id=port['id']) [ 710.092178] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] nova.exception.PortBindingFailed: Binding failed for port df4144c7-1e44-42cf-a3fe-289f5783650d, please check neutron logs for more information. [ 710.092178] env[61970]: ERROR nova.compute.manager [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] [ 710.092500] env[61970]: DEBUG nova.compute.utils [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Binding failed for port df4144c7-1e44-42cf-a3fe-289f5783650d, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 710.093219] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.973s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.095914] env[61970]: DEBUG nova.compute.manager [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Build of instance b205db9a-8170-4c53-8212-f3fc2ec11d1c was re-scheduled: Binding failed for port df4144c7-1e44-42cf-a3fe-289f5783650d, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 710.096360] env[61970]: DEBUG nova.compute.manager [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 710.096574] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Acquiring lock "refresh_cache-b205db9a-8170-4c53-8212-f3fc2ec11d1c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.096719] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Acquired lock "refresh_cache-b205db9a-8170-4c53-8212-f3fc2ec11d1c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.096878] env[61970]: DEBUG nova.network.neutron [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 710.261310] env[61970]: DEBUG oslo_concurrency.lockutils [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.373939] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355515, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.386730] env[61970]: DEBUG nova.compute.manager [req-af5fc420-ca7f-450d-9831-63b8091f0584 req-0d4be6ef-b41f-4229-b44f-3d761fd3ce08 service nova] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Received event network-changed-b758644c-56b6-4a94-b6e6-0d038fd951b4 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 710.386730] env[61970]: DEBUG nova.compute.manager [req-af5fc420-ca7f-450d-9831-63b8091f0584 req-0d4be6ef-b41f-4229-b44f-3d761fd3ce08 service nova] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Refreshing instance network info cache due to event network-changed-b758644c-56b6-4a94-b6e6-0d038fd951b4. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 710.386915] env[61970]: DEBUG oslo_concurrency.lockutils [req-af5fc420-ca7f-450d-9831-63b8091f0584 req-0d4be6ef-b41f-4229-b44f-3d761fd3ce08 service nova] Acquiring lock "refresh_cache-630145f3-cd01-41fe-b9c5-7ebf44b1e6d9" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.661452] env[61970]: DEBUG nova.network.neutron [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.663706] env[61970]: DEBUG nova.network.neutron [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.741821] env[61970]: DEBUG nova.network.neutron [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.779802] env[61970]: DEBUG nova.network.neutron [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.883364] env[61970]: DEBUG oslo_vmware.api [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355515, 'name': PowerOnVM_Task, 'duration_secs': 0.68584} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.885747] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 710.885959] env[61970]: DEBUG nova.compute.manager [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 710.886938] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fdb535d-8a56-457b-b04d-f09132d89bb5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.076581] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c445ee63-0d73-446b-afa4-41e74f2053e0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.084796] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a81863a-b7f1-4b8a-98d3-d60b0bf6152b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.114118] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4badd3c-51c6-4cb3-af0c-3c98c8d2cb21 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.121058] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0886cfe5-adc9-40d5-be99-c2c565045be6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.134655] env[61970]: DEBUG nova.compute.provider_tree [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.245381] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Releasing lock "refresh_cache-b205db9a-8170-4c53-8212-f3fc2ec11d1c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.245381] env[61970]: DEBUG nova.compute.manager [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 711.245539] env[61970]: DEBUG nova.compute.manager [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 711.245591] env[61970]: DEBUG nova.network.neutron [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 711.268248] env[61970]: DEBUG nova.network.neutron [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.288384] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Releasing lock "refresh_cache-630145f3-cd01-41fe-b9c5-7ebf44b1e6d9" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.288798] env[61970]: DEBUG nova.compute.manager [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 711.289080] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 711.289384] env[61970]: DEBUG oslo_concurrency.lockutils [req-af5fc420-ca7f-450d-9831-63b8091f0584 req-0d4be6ef-b41f-4229-b44f-3d761fd3ce08 service nova] Acquired lock "refresh_cache-630145f3-cd01-41fe-b9c5-7ebf44b1e6d9" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.289551] env[61970]: DEBUG nova.network.neutron [req-af5fc420-ca7f-450d-9831-63b8091f0584 req-0d4be6ef-b41f-4229-b44f-3d761fd3ce08 service nova] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Refreshing network info cache for port b758644c-56b6-4a94-b6e6-0d038fd951b4 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 711.290559] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4ac163fb-fdb5-4b16-8301-6e26d8bca602 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.299689] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e35b838-6dd4-4af0-9caf-1bbfcc0197fa {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.326747] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9 could not be found. [ 711.326996] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 711.327434] env[61970]: INFO nova.compute.manager [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 711.327434] env[61970]: DEBUG oslo.service.loopingcall [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 711.327794] env[61970]: DEBUG nova.compute.manager [-] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 711.327897] env[61970]: DEBUG nova.network.neutron [-] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 711.343212] env[61970]: DEBUG nova.network.neutron [-] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.405640] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.637925] env[61970]: DEBUG nova.scheduler.client.report [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 711.770885] env[61970]: DEBUG nova.network.neutron [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.807651] env[61970]: DEBUG nova.network.neutron [req-af5fc420-ca7f-450d-9831-63b8091f0584 req-0d4be6ef-b41f-4229-b44f-3d761fd3ce08 service nova] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.828583] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Acquiring lock "9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.828855] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Lock "9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.829147] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Acquiring lock "9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.829341] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Lock "9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.829508] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Lock "9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.831731] env[61970]: INFO nova.compute.manager [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Terminating instance [ 711.833463] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Acquiring lock "refresh_cache-9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.833622] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Acquired lock "refresh_cache-9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.833784] env[61970]: DEBUG nova.network.neutron [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 711.845682] env[61970]: DEBUG nova.network.neutron [-] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.911189] env[61970]: DEBUG nova.network.neutron [req-af5fc420-ca7f-450d-9831-63b8091f0584 req-0d4be6ef-b41f-4229-b44f-3d761fd3ce08 service nova] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.970537] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.970754] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.142759] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.049s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.143400] env[61970]: ERROR nova.compute.manager [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d332b101-f84b-4534-afc0-f68dbec9462c, please check neutron logs for more information. [ 712.143400] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Traceback (most recent call last): [ 712.143400] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.143400] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] self.driver.spawn(context, instance, image_meta, [ 712.143400] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 712.143400] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.143400] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.143400] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] vm_ref = self.build_virtual_machine(instance, [ 712.143400] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.143400] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.143400] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.143822] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] for vif in network_info: [ 712.143822] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 712.143822] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] return self._sync_wrapper(fn, *args, **kwargs) [ 712.143822] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 712.143822] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] self.wait() [ 712.143822] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 712.143822] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] self[:] = self._gt.wait() [ 712.143822] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.143822] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] return self._exit_event.wait() [ 712.143822] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 712.143822] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] result = hub.switch() [ 712.143822] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 712.143822] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] return self.greenlet.switch() [ 712.144213] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.144213] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] result = function(*args, **kwargs) [ 712.144213] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 712.144213] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] return func(*args, **kwargs) [ 712.144213] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.144213] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] raise e [ 712.144213] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.144213] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] nwinfo = self.network_api.allocate_for_instance( [ 712.144213] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.144213] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] created_port_ids = self._update_ports_for_instance( [ 712.144213] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.144213] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] with excutils.save_and_reraise_exception(): [ 712.144213] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.144547] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] self.force_reraise() [ 712.144547] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.144547] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] raise self.value [ 712.144547] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.144547] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] updated_port = self._update_port( [ 712.144547] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.144547] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] _ensure_no_port_binding_failure(port) [ 712.144547] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.144547] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] raise exception.PortBindingFailed(port_id=port['id']) [ 712.144547] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] nova.exception.PortBindingFailed: Binding failed for port d332b101-f84b-4534-afc0-f68dbec9462c, please check neutron logs for more information. [ 712.144547] env[61970]: ERROR nova.compute.manager [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] [ 712.144885] env[61970]: DEBUG nova.compute.utils [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Binding failed for port d332b101-f84b-4534-afc0-f68dbec9462c, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 712.145722] env[61970]: DEBUG oslo_concurrency.lockutils [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.073s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.147290] env[61970]: INFO nova.compute.claims [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.149690] env[61970]: DEBUG nova.compute.manager [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Build of instance 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb was re-scheduled: Binding failed for port d332b101-f84b-4534-afc0-f68dbec9462c, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 712.150251] env[61970]: DEBUG nova.compute.manager [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 712.150362] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Acquiring lock "refresh_cache-84c6fdb9-0858-4f25-bdc8-474ffb9f6abb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.150526] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Acquired lock "refresh_cache-84c6fdb9-0858-4f25-bdc8-474ffb9f6abb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.150691] env[61970]: DEBUG nova.network.neutron [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 712.273704] env[61970]: INFO nova.compute.manager [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] [instance: b205db9a-8170-4c53-8212-f3fc2ec11d1c] Took 1.03 seconds to deallocate network for instance. [ 712.347440] env[61970]: INFO nova.compute.manager [-] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Took 1.02 seconds to deallocate network for instance. [ 712.349456] env[61970]: DEBUG nova.compute.claims [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 712.349634] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.352389] env[61970]: DEBUG nova.network.neutron [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.403630] env[61970]: DEBUG nova.network.neutron [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.416921] env[61970]: DEBUG oslo_concurrency.lockutils [req-af5fc420-ca7f-450d-9831-63b8091f0584 req-0d4be6ef-b41f-4229-b44f-3d761fd3ce08 service nova] Releasing lock "refresh_cache-630145f3-cd01-41fe-b9c5-7ebf44b1e6d9" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.417120] env[61970]: DEBUG nova.compute.manager [req-af5fc420-ca7f-450d-9831-63b8091f0584 req-0d4be6ef-b41f-4229-b44f-3d761fd3ce08 service nova] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Received event network-vif-deleted-b758644c-56b6-4a94-b6e6-0d038fd951b4 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 712.673509] env[61970]: DEBUG nova.network.neutron [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.740090] env[61970]: DEBUG nova.network.neutron [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.906643] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Releasing lock "refresh_cache-9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.907094] env[61970]: DEBUG nova.compute.manager [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 712.907298] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 712.908221] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b93f6d8-57b9-45d6-b6c0-0724856d1e38 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.917059] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 712.917316] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00bf7367-8b53-4f55-b7af-c8e4aa4de8a1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.923945] env[61970]: DEBUG oslo_vmware.api [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 712.923945] env[61970]: value = "task-1355516" [ 712.923945] env[61970]: _type = "Task" [ 712.923945] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.933254] env[61970]: DEBUG oslo_vmware.api [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355516, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.242545] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Releasing lock "refresh_cache-84c6fdb9-0858-4f25-bdc8-474ffb9f6abb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.242781] env[61970]: DEBUG nova.compute.manager [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 713.242961] env[61970]: DEBUG nova.compute.manager [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 713.243148] env[61970]: DEBUG nova.network.neutron [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 713.263482] env[61970]: DEBUG nova.network.neutron [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.300615] env[61970]: INFO nova.scheduler.client.report [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Deleted allocations for instance b205db9a-8170-4c53-8212-f3fc2ec11d1c [ 713.433363] env[61970]: DEBUG oslo_vmware.api [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355516, 'name': PowerOffVM_Task, 'duration_secs': 0.111061} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.433544] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 713.433712] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 713.434018] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-015ed63b-1ece-46ea-a9ab-e7c1396e0314 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.457562] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 713.458228] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 713.458228] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Deleting the datastore file [datastore2] 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 713.458228] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a6aed1d7-9c0b-42f3-9d38-5488a1c43e7e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.465012] env[61970]: DEBUG oslo_vmware.api [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for the task: (returnval){ [ 713.465012] env[61970]: value = "task-1355518" [ 713.465012] env[61970]: _type = "Task" [ 713.465012] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.473830] env[61970]: DEBUG oslo_vmware.api [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355518, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.532924] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f762d41-1a3b-47a6-ba2d-14696b0909e6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.540098] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb8e606f-63e9-43ba-9f49-3bc5dba84e64 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.570525] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8dfbc8f-5fad-4b6f-a1a0-2bda9e2637a8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.577507] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae3ab25-914a-4c50-b607-2b3f2c729c07 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.592342] env[61970]: DEBUG nova.compute.provider_tree [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.766508] env[61970]: DEBUG nova.network.neutron [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.809820] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8ec77b33-e3c5-4e35-9796-566559bc705b tempest-ServerDiagnosticsTest-1487417783 tempest-ServerDiagnosticsTest-1487417783-project-member] Lock "b205db9a-8170-4c53-8212-f3fc2ec11d1c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.531s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.975130] env[61970]: DEBUG oslo_vmware.api [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Task: {'id': task-1355518, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101307} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.975385] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 713.975558] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 713.975728] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 713.975892] env[61970]: INFO nova.compute.manager [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Took 1.07 seconds to destroy the instance on the hypervisor. [ 713.976143] env[61970]: DEBUG oslo.service.loopingcall [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 713.976328] env[61970]: DEBUG nova.compute.manager [-] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 713.976420] env[61970]: DEBUG nova.network.neutron [-] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 713.990703] env[61970]: DEBUG nova.network.neutron [-] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.095736] env[61970]: DEBUG nova.scheduler.client.report [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.269576] env[61970]: INFO nova.compute.manager [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] [instance: 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb] Took 1.03 seconds to deallocate network for instance. [ 714.312829] env[61970]: DEBUG nova.compute.manager [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 714.493534] env[61970]: DEBUG nova.network.neutron [-] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.601250] env[61970]: DEBUG oslo_concurrency.lockutils [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.601640] env[61970]: DEBUG nova.compute.manager [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 714.604100] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.081s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.837195] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.996517] env[61970]: INFO nova.compute.manager [-] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Took 1.02 seconds to deallocate network for instance. [ 715.112235] env[61970]: DEBUG nova.compute.utils [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 715.114459] env[61970]: DEBUG nova.compute.manager [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Not allocating networking since 'none' was specified. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 715.296708] env[61970]: INFO nova.scheduler.client.report [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Deleted allocations for instance 84c6fdb9-0858-4f25-bdc8-474ffb9f6abb [ 715.506111] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.515994] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4fe7c1-4eb6-4452-a717-afa11a48132b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.523890] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5500bf04-fea4-4b0a-b0be-a3cab7abd7c1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.553507] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a639256-dfc4-47d2-b82c-6531d3f12929 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.560639] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dbec549-de10-4c07-8663-1aa15880df9b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.573580] env[61970]: DEBUG nova.compute.provider_tree [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.614676] env[61970]: DEBUG nova.compute.manager [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 715.806568] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4547aea3-6b3a-4306-8716-38c77fdeaec7 tempest-ServerRescueTestJSON-2107792598 tempest-ServerRescueTestJSON-2107792598-project-member] Lock "84c6fdb9-0858-4f25-bdc8-474ffb9f6abb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.325s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.078465] env[61970]: DEBUG nova.scheduler.client.report [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.309758] env[61970]: DEBUG nova.compute.manager [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 716.583973] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.980s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.584683] env[61970]: ERROR nova.compute.manager [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3639de38-74f5-4cd0-ae6e-223790e16f20, please check neutron logs for more information. [ 716.584683] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Traceback (most recent call last): [ 716.584683] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 716.584683] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] self.driver.spawn(context, instance, image_meta, [ 716.584683] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 716.584683] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 716.584683] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 716.584683] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] vm_ref = self.build_virtual_machine(instance, [ 716.584683] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 716.584683] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] vif_infos = vmwarevif.get_vif_info(self._session, [ 716.584683] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 716.585081] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] for vif in network_info: [ 716.585081] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 716.585081] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] return self._sync_wrapper(fn, *args, **kwargs) [ 716.585081] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 716.585081] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] self.wait() [ 716.585081] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 716.585081] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] self[:] = self._gt.wait() [ 716.585081] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 716.585081] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] return self._exit_event.wait() [ 716.585081] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 716.585081] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] current.throw(*self._exc) [ 716.585081] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.585081] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] result = function(*args, **kwargs) [ 716.585414] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 716.585414] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] return func(*args, **kwargs) [ 716.585414] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.585414] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] raise e [ 716.585414] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.585414] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] nwinfo = self.network_api.allocate_for_instance( [ 716.585414] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.585414] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] created_port_ids = self._update_ports_for_instance( [ 716.585414] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.585414] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] with excutils.save_and_reraise_exception(): [ 716.585414] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.585414] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] self.force_reraise() [ 716.585414] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.585862] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] raise self.value [ 716.585862] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.585862] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] updated_port = self._update_port( [ 716.585862] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.585862] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] _ensure_no_port_binding_failure(port) [ 716.585862] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.585862] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] raise exception.PortBindingFailed(port_id=port['id']) [ 716.585862] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] nova.exception.PortBindingFailed: Binding failed for port 3639de38-74f5-4cd0-ae6e-223790e16f20, please check neutron logs for more information. [ 716.585862] env[61970]: ERROR nova.compute.manager [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] [ 716.585862] env[61970]: DEBUG nova.compute.utils [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Binding failed for port 3639de38-74f5-4cd0-ae6e-223790e16f20, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 716.586653] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.013s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.588399] env[61970]: INFO nova.compute.claims [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 716.590940] env[61970]: DEBUG nova.compute.manager [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Build of instance 3661a95f-a75c-4f0f-ad25-428afdb7bccb was re-scheduled: Binding failed for port 3639de38-74f5-4cd0-ae6e-223790e16f20, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 716.591475] env[61970]: DEBUG nova.compute.manager [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 716.591636] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Acquiring lock "refresh_cache-3661a95f-a75c-4f0f-ad25-428afdb7bccb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.591782] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Acquired lock "refresh_cache-3661a95f-a75c-4f0f-ad25-428afdb7bccb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.591939] env[61970]: DEBUG nova.network.neutron [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 716.631478] env[61970]: DEBUG nova.compute.manager [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 716.664098] env[61970]: DEBUG nova.virt.hardware [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 716.664453] env[61970]: DEBUG nova.virt.hardware [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 716.665054] env[61970]: DEBUG nova.virt.hardware [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 716.665158] env[61970]: DEBUG nova.virt.hardware [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 716.665298] env[61970]: DEBUG nova.virt.hardware [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 716.665485] env[61970]: DEBUG nova.virt.hardware [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 716.665857] env[61970]: DEBUG nova.virt.hardware [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 716.666139] env[61970]: DEBUG nova.virt.hardware [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 716.666363] env[61970]: DEBUG nova.virt.hardware [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 716.666565] env[61970]: DEBUG nova.virt.hardware [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 716.666777] env[61970]: DEBUG nova.virt.hardware [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 716.667942] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cea3597-ab8d-44f7-99b5-5271fae565fb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.677560] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd9040de-13c6-4afe-ad09-48c3c7790c41 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.694068] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Instance VIF info [] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 716.700256] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Creating folder: Project (6759429a3e1b4b46a3a45166a0589ddf). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 716.700442] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3857ba3a-206b-4154-8f16-abc9f4e6a58c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.711010] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Created folder: Project (6759429a3e1b4b46a3a45166a0589ddf) in parent group-v288740. [ 716.711224] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Creating folder: Instances. Parent ref: group-v288756. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 716.711455] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-47898fce-8eaa-4be2-afc0-32d23cdbed54 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.720440] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Created folder: Instances in parent group-v288756. [ 716.720682] env[61970]: DEBUG oslo.service.loopingcall [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 716.720907] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 716.721180] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-307f0ba9-9a05-458f-9090-64c6e0cf73ab {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.738205] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 716.738205] env[61970]: value = "task-1355521" [ 716.738205] env[61970]: _type = "Task" [ 716.738205] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.745701] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355521, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.838845] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.119855] env[61970]: DEBUG nova.network.neutron [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.250928] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355521, 'name': CreateVM_Task, 'duration_secs': 0.270151} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.251109] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 717.251526] env[61970]: DEBUG oslo_concurrency.lockutils [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.251702] env[61970]: DEBUG oslo_concurrency.lockutils [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.253802] env[61970]: DEBUG oslo_concurrency.lockutils [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 717.253802] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d360b04-b40d-49dc-b135-678f92b48eed {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.254528] env[61970]: DEBUG nova.network.neutron [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.258943] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 717.258943] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b3a7d5-fb53-942a-5d32-d8f49d09ffc7" [ 717.258943] env[61970]: _type = "Task" [ 717.258943] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.267235] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b3a7d5-fb53-942a-5d32-d8f49d09ffc7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.757351] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Releasing lock "refresh_cache-3661a95f-a75c-4f0f-ad25-428afdb7bccb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.757587] env[61970]: DEBUG nova.compute.manager [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 717.757770] env[61970]: DEBUG nova.compute.manager [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 717.757935] env[61970]: DEBUG nova.network.neutron [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 717.775259] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b3a7d5-fb53-942a-5d32-d8f49d09ffc7, 'name': SearchDatastore_Task, 'duration_secs': 0.009057} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.778830] env[61970]: DEBUG oslo_concurrency.lockutils [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.778830] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 717.778830] env[61970]: DEBUG oslo_concurrency.lockutils [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.778830] env[61970]: DEBUG oslo_concurrency.lockutils [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.779036] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 717.779036] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ae49e688-1988-42de-a3bc-23ad3e02cec2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.783305] env[61970]: DEBUG nova.network.neutron [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.786580] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 717.786741] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 717.787665] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13d6dfbc-9a72-43ee-86b0-c3ac9e5e31c8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.795452] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 717.795452] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5250b812-645d-ff68-e187-6ab4a77b072a" [ 717.795452] env[61970]: _type = "Task" [ 717.795452] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.803531] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5250b812-645d-ff68-e187-6ab4a77b072a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.053678] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50cec660-94ed-4f68-8de3-a37f81ea2eba {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.061633] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f3f75a3-07fb-4b2d-a613-fa0d72065ffb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.093347] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77244bd-db9b-45ee-ae6f-01a4e621744e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.101029] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d41364f-606e-4f60-82a5-778cca8231c4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.113728] env[61970]: DEBUG nova.compute.provider_tree [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.289323] env[61970]: DEBUG nova.network.neutron [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.307400] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5250b812-645d-ff68-e187-6ab4a77b072a, 'name': SearchDatastore_Task, 'duration_secs': 0.009283} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.308217] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f385656b-b600-4e08-851e-26fd4da53f88 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.313338] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 718.313338] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529e135d-e046-c238-b6df-3fcfbd7b6486" [ 718.313338] env[61970]: _type = "Task" [ 718.313338] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.321209] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529e135d-e046-c238-b6df-3fcfbd7b6486, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.616898] env[61970]: DEBUG nova.scheduler.client.report [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 718.792742] env[61970]: INFO nova.compute.manager [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] [instance: 3661a95f-a75c-4f0f-ad25-428afdb7bccb] Took 1.03 seconds to deallocate network for instance. [ 718.827358] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529e135d-e046-c238-b6df-3fcfbd7b6486, 'name': SearchDatastore_Task, 'duration_secs': 0.008459} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.827762] env[61970]: DEBUG oslo_concurrency.lockutils [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.830303] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 915c3b6c-640b-430f-b264-40bf85c642b9/915c3b6c-640b-430f-b264-40bf85c642b9.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 718.830303] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-56d31512-8e71-483a-a028-336d1ad6ebc7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.836030] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 718.836030] env[61970]: value = "task-1355522" [ 718.836030] env[61970]: _type = "Task" [ 718.836030] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.846883] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355522, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.122736] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.536s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.123412] env[61970]: DEBUG nova.compute.manager [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 719.127302] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.761s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.345164] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355522, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.443882} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.345497] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 915c3b6c-640b-430f-b264-40bf85c642b9/915c3b6c-640b-430f-b264-40bf85c642b9.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 719.345659] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 719.345896] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a398838-c93c-420f-af7b-1a3b1e405a35 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.352692] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 719.352692] env[61970]: value = "task-1355523" [ 719.352692] env[61970]: _type = "Task" [ 719.352692] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.359808] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355523, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.628766] env[61970]: DEBUG nova.compute.utils [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 719.630161] env[61970]: DEBUG nova.compute.manager [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 719.630339] env[61970]: DEBUG nova.network.neutron [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 719.695044] env[61970]: DEBUG nova.policy [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '40b50e68eabb475bb031cb63cb036d65', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cff16a8a40ed4c32b41dd81197c2f8c0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 719.820878] env[61970]: INFO nova.scheduler.client.report [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Deleted allocations for instance 3661a95f-a75c-4f0f-ad25-428afdb7bccb [ 719.863997] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355523, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059698} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.864149] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 719.865033] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb6575c-3f58-415b-aa1a-8787fae6cfa7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.884173] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] 915c3b6c-640b-430f-b264-40bf85c642b9/915c3b6c-640b-430f-b264-40bf85c642b9.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 719.886951] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1ee7a7a-a90a-46c5-8ba1-d74ab8deacdc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.906220] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 719.906220] env[61970]: value = "task-1355524" [ 719.906220] env[61970]: _type = "Task" [ 719.906220] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.915744] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355524, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.040992] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aecb94fd-6b6e-405a-844c-4146443a174b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.049820] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe2a735-3086-4694-a257-077a1acdc8ae {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.090025] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ae1366-ffe0-4420-a21c-09912852d0a9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.097702] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f03efc-7d3a-4e47-bced-a335a2e01969 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.111279] env[61970]: DEBUG nova.compute.provider_tree [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.134353] env[61970]: DEBUG nova.compute.manager [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 720.257461] env[61970]: DEBUG nova.network.neutron [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Successfully created port: 65286ee8-9e57-467d-887e-f5cfacaa25e5 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 720.331862] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4700b89e-2b7e-4e4e-82f3-f037bf20de5e tempest-AttachInterfacesUnderV243Test-308532418 tempest-AttachInterfacesUnderV243Test-308532418-project-member] Lock "3661a95f-a75c-4f0f-ad25-428afdb7bccb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.740s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.416463] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355524, 'name': ReconfigVM_Task, 'duration_secs': 0.274365} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.416785] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Reconfigured VM instance instance-00000027 to attach disk [datastore2] 915c3b6c-640b-430f-b264-40bf85c642b9/915c3b6c-640b-430f-b264-40bf85c642b9.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 720.417349] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-41fc5243-07dc-4df9-abdb-3cda186e47c9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.423341] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 720.423341] env[61970]: value = "task-1355525" [ 720.423341] env[61970]: _type = "Task" [ 720.423341] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.431294] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355525, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.614017] env[61970]: DEBUG nova.scheduler.client.report [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.834479] env[61970]: DEBUG nova.compute.manager [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 720.933100] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355525, 'name': Rename_Task, 'duration_secs': 0.166642} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.933376] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 720.933612] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4493022f-de8a-49b2-a1fa-3fe69fecd8d8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.939935] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 720.939935] env[61970]: value = "task-1355526" [ 720.939935] env[61970]: _type = "Task" [ 720.939935] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.952022] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355526, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.055824] env[61970]: DEBUG nova.compute.manager [req-229383ac-5f14-43f8-8ba6-6d0210279445 req-adfcdcab-6543-4ec8-9d8c-bfaead595d31 service nova] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Received event network-changed-65286ee8-9e57-467d-887e-f5cfacaa25e5 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 721.056255] env[61970]: DEBUG nova.compute.manager [req-229383ac-5f14-43f8-8ba6-6d0210279445 req-adfcdcab-6543-4ec8-9d8c-bfaead595d31 service nova] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Refreshing instance network info cache due to event network-changed-65286ee8-9e57-467d-887e-f5cfacaa25e5. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 721.056786] env[61970]: DEBUG oslo_concurrency.lockutils [req-229383ac-5f14-43f8-8ba6-6d0210279445 req-adfcdcab-6543-4ec8-9d8c-bfaead595d31 service nova] Acquiring lock "refresh_cache-1a9257f6-5493-4ebb-8c37-88e11480a4f6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.056786] env[61970]: DEBUG oslo_concurrency.lockutils [req-229383ac-5f14-43f8-8ba6-6d0210279445 req-adfcdcab-6543-4ec8-9d8c-bfaead595d31 service nova] Acquired lock "refresh_cache-1a9257f6-5493-4ebb-8c37-88e11480a4f6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.056786] env[61970]: DEBUG nova.network.neutron [req-229383ac-5f14-43f8-8ba6-6d0210279445 req-adfcdcab-6543-4ec8-9d8c-bfaead595d31 service nova] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Refreshing network info cache for port 65286ee8-9e57-467d-887e-f5cfacaa25e5 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 721.125295] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.996s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.125295] env[61970]: ERROR nova.compute.manager [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ce17d99a-8184-48b2-a9ac-2afbf5f36def, please check neutron logs for more information. [ 721.125295] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Traceback (most recent call last): [ 721.125295] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 721.125295] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] self.driver.spawn(context, instance, image_meta, [ 721.125295] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 721.125295] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] self._vmops.spawn(context, instance, image_meta, injected_files, [ 721.125295] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 721.125295] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] vm_ref = self.build_virtual_machine(instance, [ 721.125603] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 721.125603] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] vif_infos = vmwarevif.get_vif_info(self._session, [ 721.125603] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 721.125603] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] for vif in network_info: [ 721.125603] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 721.125603] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] return self._sync_wrapper(fn, *args, **kwargs) [ 721.125603] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 721.125603] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] self.wait() [ 721.125603] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 721.125603] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] self[:] = self._gt.wait() [ 721.125603] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 721.125603] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] return self._exit_event.wait() [ 721.125603] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 721.125989] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] result = hub.switch() [ 721.125989] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 721.125989] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] return self.greenlet.switch() [ 721.125989] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.125989] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] result = function(*args, **kwargs) [ 721.125989] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 721.125989] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] return func(*args, **kwargs) [ 721.125989] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.125989] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] raise e [ 721.125989] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.125989] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] nwinfo = self.network_api.allocate_for_instance( [ 721.125989] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.125989] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] created_port_ids = self._update_ports_for_instance( [ 721.126487] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.126487] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] with excutils.save_and_reraise_exception(): [ 721.126487] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.126487] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] self.force_reraise() [ 721.126487] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.126487] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] raise self.value [ 721.126487] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.126487] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] updated_port = self._update_port( [ 721.126487] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.126487] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] _ensure_no_port_binding_failure(port) [ 721.126487] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.126487] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] raise exception.PortBindingFailed(port_id=port['id']) [ 721.126846] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] nova.exception.PortBindingFailed: Binding failed for port ce17d99a-8184-48b2-a9ac-2afbf5f36def, please check neutron logs for more information. [ 721.126846] env[61970]: ERROR nova.compute.manager [instance: 7dda08a4-83ae-4882-9469-324d948c9749] [ 721.126846] env[61970]: DEBUG nova.compute.utils [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Binding failed for port ce17d99a-8184-48b2-a9ac-2afbf5f36def, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 721.126846] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.413s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.131164] env[61970]: INFO nova.compute.claims [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 721.132729] env[61970]: DEBUG nova.compute.manager [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Build of instance 7dda08a4-83ae-4882-9469-324d948c9749 was re-scheduled: Binding failed for port ce17d99a-8184-48b2-a9ac-2afbf5f36def, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 721.132953] env[61970]: DEBUG nova.compute.manager [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 721.133275] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Acquiring lock "refresh_cache-7dda08a4-83ae-4882-9469-324d948c9749" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.133392] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Acquired lock "refresh_cache-7dda08a4-83ae-4882-9469-324d948c9749" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.133775] env[61970]: DEBUG nova.network.neutron [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 721.143150] env[61970]: DEBUG nova.compute.manager [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 721.176231] env[61970]: DEBUG nova.virt.hardware [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 721.176480] env[61970]: DEBUG nova.virt.hardware [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 721.176632] env[61970]: DEBUG nova.virt.hardware [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 721.176809] env[61970]: DEBUG nova.virt.hardware [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 721.176950] env[61970]: DEBUG nova.virt.hardware [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 721.177264] env[61970]: DEBUG nova.virt.hardware [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 721.177497] env[61970]: DEBUG nova.virt.hardware [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 721.177657] env[61970]: DEBUG nova.virt.hardware [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 721.177820] env[61970]: DEBUG nova.virt.hardware [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 721.177980] env[61970]: DEBUG nova.virt.hardware [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 721.178166] env[61970]: DEBUG nova.virt.hardware [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 721.179338] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e0cc30-1299-4095-b86c-562d9b54d116 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.187814] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a9978b3-27d0-4d52-976a-33d8e0d57b61 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.290235] env[61970]: ERROR nova.compute.manager [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 65286ee8-9e57-467d-887e-f5cfacaa25e5, please check neutron logs for more information. [ 721.290235] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 721.290235] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.290235] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 721.290235] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.290235] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 721.290235] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.290235] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 721.290235] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.290235] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 721.290235] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.290235] env[61970]: ERROR nova.compute.manager raise self.value [ 721.290235] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.290235] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 721.290235] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.290235] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 721.290696] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.290696] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 721.290696] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 65286ee8-9e57-467d-887e-f5cfacaa25e5, please check neutron logs for more information. [ 721.290696] env[61970]: ERROR nova.compute.manager [ 721.290696] env[61970]: Traceback (most recent call last): [ 721.290696] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 721.290696] env[61970]: listener.cb(fileno) [ 721.290696] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.290696] env[61970]: result = function(*args, **kwargs) [ 721.290696] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 721.290696] env[61970]: return func(*args, **kwargs) [ 721.290696] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.290696] env[61970]: raise e [ 721.290696] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.290696] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 721.290696] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.290696] env[61970]: created_port_ids = self._update_ports_for_instance( [ 721.290696] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.290696] env[61970]: with excutils.save_and_reraise_exception(): [ 721.290696] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.290696] env[61970]: self.force_reraise() [ 721.290696] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.290696] env[61970]: raise self.value [ 721.290696] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.290696] env[61970]: updated_port = self._update_port( [ 721.290696] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.290696] env[61970]: _ensure_no_port_binding_failure(port) [ 721.290696] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.290696] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 721.291794] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 65286ee8-9e57-467d-887e-f5cfacaa25e5, please check neutron logs for more information. [ 721.291794] env[61970]: Removing descriptor: 17 [ 721.291794] env[61970]: ERROR nova.compute.manager [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 65286ee8-9e57-467d-887e-f5cfacaa25e5, please check neutron logs for more information. [ 721.291794] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Traceback (most recent call last): [ 721.291794] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 721.291794] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] yield resources [ 721.291794] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 721.291794] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] self.driver.spawn(context, instance, image_meta, [ 721.291794] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 721.291794] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 721.291794] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 721.291794] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] vm_ref = self.build_virtual_machine(instance, [ 721.292109] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 721.292109] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 721.292109] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 721.292109] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] for vif in network_info: [ 721.292109] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 721.292109] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] return self._sync_wrapper(fn, *args, **kwargs) [ 721.292109] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 721.292109] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] self.wait() [ 721.292109] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 721.292109] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] self[:] = self._gt.wait() [ 721.292109] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 721.292109] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] return self._exit_event.wait() [ 721.292109] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 721.292507] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] result = hub.switch() [ 721.292507] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 721.292507] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] return self.greenlet.switch() [ 721.292507] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.292507] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] result = function(*args, **kwargs) [ 721.292507] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 721.292507] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] return func(*args, **kwargs) [ 721.292507] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.292507] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] raise e [ 721.292507] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.292507] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] nwinfo = self.network_api.allocate_for_instance( [ 721.292507] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.292507] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] created_port_ids = self._update_ports_for_instance( [ 721.292958] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.292958] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] with excutils.save_and_reraise_exception(): [ 721.292958] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.292958] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] self.force_reraise() [ 721.292958] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.292958] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] raise self.value [ 721.292958] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.292958] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] updated_port = self._update_port( [ 721.292958] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.292958] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] _ensure_no_port_binding_failure(port) [ 721.292958] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.292958] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] raise exception.PortBindingFailed(port_id=port['id']) [ 721.293325] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] nova.exception.PortBindingFailed: Binding failed for port 65286ee8-9e57-467d-887e-f5cfacaa25e5, please check neutron logs for more information. [ 721.293325] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] [ 721.293325] env[61970]: INFO nova.compute.manager [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Terminating instance [ 721.293325] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Acquiring lock "refresh_cache-1a9257f6-5493-4ebb-8c37-88e11480a4f6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.355616] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.454211] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355526, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.581227] env[61970]: DEBUG nova.network.neutron [req-229383ac-5f14-43f8-8ba6-6d0210279445 req-adfcdcab-6543-4ec8-9d8c-bfaead595d31 service nova] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.630866] env[61970]: DEBUG nova.network.neutron [req-229383ac-5f14-43f8-8ba6-6d0210279445 req-adfcdcab-6543-4ec8-9d8c-bfaead595d31 service nova] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.655121] env[61970]: DEBUG nova.network.neutron [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.754320] env[61970]: DEBUG nova.network.neutron [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.954022] env[61970]: DEBUG oslo_vmware.api [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355526, 'name': PowerOnVM_Task, 'duration_secs': 0.599538} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.954022] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 721.954022] env[61970]: INFO nova.compute.manager [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Took 5.32 seconds to spawn the instance on the hypervisor. [ 721.954022] env[61970]: DEBUG nova.compute.manager [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 721.958022] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05bec6b-fbea-451d-b336-0b41035b9955 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.133415] env[61970]: DEBUG oslo_concurrency.lockutils [req-229383ac-5f14-43f8-8ba6-6d0210279445 req-adfcdcab-6543-4ec8-9d8c-bfaead595d31 service nova] Releasing lock "refresh_cache-1a9257f6-5493-4ebb-8c37-88e11480a4f6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.133842] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Acquired lock "refresh_cache-1a9257f6-5493-4ebb-8c37-88e11480a4f6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.134105] env[61970]: DEBUG nova.network.neutron [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 722.257684] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Releasing lock "refresh_cache-7dda08a4-83ae-4882-9469-324d948c9749" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.257684] env[61970]: DEBUG nova.compute.manager [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 722.257684] env[61970]: DEBUG nova.compute.manager [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.257684] env[61970]: DEBUG nova.network.neutron [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 722.283514] env[61970]: DEBUG nova.network.neutron [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.476823] env[61970]: INFO nova.compute.manager [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Took 26.42 seconds to build instance. [ 722.594060] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-586d3bc8-5c08-4080-aee5-fbca8d036a89 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.601332] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda9b68b-fdd1-438b-b299-1070ee14c950 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.633908] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36828806-64c7-4cc5-b5af-752530f276d9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.647857] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb85aa1e-1210-4dea-91db-7c74a8ee086c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.662551] env[61970]: DEBUG nova.compute.provider_tree [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.671102] env[61970]: DEBUG nova.network.neutron [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.738582] env[61970]: DEBUG nova.network.neutron [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.788328] env[61970]: DEBUG nova.network.neutron [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.983017] env[61970]: DEBUG oslo_concurrency.lockutils [None req-74b15b2e-2c78-46e9-8e93-57002ea0b459 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Lock "915c3b6c-640b-430f-b264-40bf85c642b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.946s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.080958] env[61970]: DEBUG nova.compute.manager [req-64927d1f-020d-42a8-8e7c-44847321ef01 req-afbb75b8-be41-44f6-8bac-060a5f4c3425 service nova] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Received event network-vif-deleted-65286ee8-9e57-467d-887e-f5cfacaa25e5 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 723.143196] env[61970]: INFO nova.compute.manager [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Rebuilding instance [ 723.165599] env[61970]: DEBUG nova.scheduler.client.report [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.190938] env[61970]: DEBUG nova.compute.manager [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 723.191824] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43e64655-4023-4e00-a5c4-908cc37c4e53 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.241598] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Releasing lock "refresh_cache-1a9257f6-5493-4ebb-8c37-88e11480a4f6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.241979] env[61970]: DEBUG nova.compute.manager [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 723.242188] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 723.242706] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d5b81d28-dbc3-420b-8ed2-da4ab951ce37 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.252664] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-467813b8-f420-432e-8cd9-7d1247430cc1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.274131] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1a9257f6-5493-4ebb-8c37-88e11480a4f6 could not be found. [ 723.274349] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 723.274528] env[61970]: INFO nova.compute.manager [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 723.274766] env[61970]: DEBUG oslo.service.loopingcall [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 723.274962] env[61970]: DEBUG nova.compute.manager [-] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 723.275068] env[61970]: DEBUG nova.network.neutron [-] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 723.289583] env[61970]: DEBUG nova.network.neutron [-] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.296743] env[61970]: INFO nova.compute.manager [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] [instance: 7dda08a4-83ae-4882-9469-324d948c9749] Took 1.04 seconds to deallocate network for instance. [ 723.485394] env[61970]: DEBUG nova.compute.manager [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 723.674604] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.548s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.675148] env[61970]: DEBUG nova.compute.manager [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 723.677782] env[61970]: DEBUG oslo_concurrency.lockutils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.052s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.703190] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 723.703494] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d491f6d-31b7-45ad-af3a-34504dd5f08b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.711371] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 723.711371] env[61970]: value = "task-1355527" [ 723.711371] env[61970]: _type = "Task" [ 723.711371] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.720107] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355527, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.791924] env[61970]: DEBUG nova.network.neutron [-] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.005568] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.182308] env[61970]: DEBUG nova.compute.utils [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 724.186930] env[61970]: DEBUG nova.compute.manager [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 724.187115] env[61970]: DEBUG nova.network.neutron [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 724.224125] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355527, 'name': PowerOffVM_Task, 'duration_secs': 0.187766} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.224459] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 724.224687] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 724.225461] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d68b47-8928-402b-80c3-9034e39f6fb5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.234285] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 724.235077] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d105892-2158-48a2-b0e6-b7a7d4ba57e5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.241729] env[61970]: DEBUG nova.policy [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bcda5eb7709443c581aa78088cb77540', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6560203e6f1e45b59463a749b03d4564', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 724.258768] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 724.258964] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 724.259155] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Deleting the datastore file [datastore2] 915c3b6c-640b-430f-b264-40bf85c642b9 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 724.259424] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6ca49833-7b3d-4168-9b58-892ebdb4acee {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.267429] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 724.267429] env[61970]: value = "task-1355529" [ 724.267429] env[61970]: _type = "Task" [ 724.267429] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.276705] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355529, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.295454] env[61970]: INFO nova.compute.manager [-] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Took 1.02 seconds to deallocate network for instance. [ 724.300597] env[61970]: DEBUG nova.compute.claims [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 724.300791] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.326370] env[61970]: INFO nova.scheduler.client.report [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Deleted allocations for instance 7dda08a4-83ae-4882-9469-324d948c9749 [ 724.556715] env[61970]: DEBUG nova.network.neutron [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Successfully created port: 091e50a0-7fd8-4d7c-9035-4a29f32819ae {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 724.625880] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55cbd90c-3d55-404f-af6d-b2dc36577f4b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.633638] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195a50e8-f29c-4b55-8bf6-06402fa77371 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.664665] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11bff112-a5a4-4377-9226-aede8dc99822 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.673390] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baf10caa-a9bc-4196-bc21-7e93d0ee4925 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.687080] env[61970]: DEBUG nova.compute.provider_tree [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.688620] env[61970]: DEBUG nova.compute.manager [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 724.776933] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355529, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094936} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.777342] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 724.777466] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 724.777648] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 724.836308] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6454a13a-c63e-4184-8180-0aa9fbe0c077 tempest-ServersTestJSON-490783360 tempest-ServersTestJSON-490783360-project-member] Lock "7dda08a4-83ae-4882-9469-324d948c9749" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.781s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.193966] env[61970]: DEBUG nova.scheduler.client.report [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.339225] env[61970]: DEBUG nova.compute.manager [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 725.448131] env[61970]: DEBUG nova.compute.manager [req-6fb19add-3e77-4c61-a624-0bd46d3c632e req-59e507a9-2fdf-460a-b7ef-4053d56443eb service nova] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Received event network-changed-091e50a0-7fd8-4d7c-9035-4a29f32819ae {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.448324] env[61970]: DEBUG nova.compute.manager [req-6fb19add-3e77-4c61-a624-0bd46d3c632e req-59e507a9-2fdf-460a-b7ef-4053d56443eb service nova] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Refreshing instance network info cache due to event network-changed-091e50a0-7fd8-4d7c-9035-4a29f32819ae. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 725.448489] env[61970]: DEBUG oslo_concurrency.lockutils [req-6fb19add-3e77-4c61-a624-0bd46d3c632e req-59e507a9-2fdf-460a-b7ef-4053d56443eb service nova] Acquiring lock "refresh_cache-762efd5e-c95e-4718-98de-2e0b05226c06" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.448631] env[61970]: DEBUG oslo_concurrency.lockutils [req-6fb19add-3e77-4c61-a624-0bd46d3c632e req-59e507a9-2fdf-460a-b7ef-4053d56443eb service nova] Acquired lock "refresh_cache-762efd5e-c95e-4718-98de-2e0b05226c06" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.448787] env[61970]: DEBUG nova.network.neutron [req-6fb19add-3e77-4c61-a624-0bd46d3c632e req-59e507a9-2fdf-460a-b7ef-4053d56443eb service nova] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Refreshing network info cache for port 091e50a0-7fd8-4d7c-9035-4a29f32819ae {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 725.608971] env[61970]: ERROR nova.compute.manager [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 091e50a0-7fd8-4d7c-9035-4a29f32819ae, please check neutron logs for more information. [ 725.608971] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 725.608971] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.608971] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 725.608971] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.608971] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 725.608971] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.608971] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 725.608971] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.608971] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 725.608971] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.608971] env[61970]: ERROR nova.compute.manager raise self.value [ 725.608971] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.608971] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 725.608971] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.608971] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 725.609710] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.609710] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 725.609710] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 091e50a0-7fd8-4d7c-9035-4a29f32819ae, please check neutron logs for more information. [ 725.609710] env[61970]: ERROR nova.compute.manager [ 725.609710] env[61970]: Traceback (most recent call last): [ 725.609710] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 725.609710] env[61970]: listener.cb(fileno) [ 725.609710] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.609710] env[61970]: result = function(*args, **kwargs) [ 725.609710] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 725.609710] env[61970]: return func(*args, **kwargs) [ 725.609710] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.609710] env[61970]: raise e [ 725.609710] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.609710] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 725.609710] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.609710] env[61970]: created_port_ids = self._update_ports_for_instance( [ 725.609710] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.609710] env[61970]: with excutils.save_and_reraise_exception(): [ 725.609710] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.609710] env[61970]: self.force_reraise() [ 725.609710] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.609710] env[61970]: raise self.value [ 725.609710] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.609710] env[61970]: updated_port = self._update_port( [ 725.609710] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.609710] env[61970]: _ensure_no_port_binding_failure(port) [ 725.609710] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.609710] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 725.610490] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 091e50a0-7fd8-4d7c-9035-4a29f32819ae, please check neutron logs for more information. [ 725.610490] env[61970]: Removing descriptor: 17 [ 725.702920] env[61970]: DEBUG oslo_concurrency.lockutils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.025s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.703560] env[61970]: ERROR nova.compute.manager [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port da878503-b70a-4e57-b4d1-eacf4bb3c3f3, please check neutron logs for more information. [ 725.703560] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] Traceback (most recent call last): [ 725.703560] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 725.703560] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] self.driver.spawn(context, instance, image_meta, [ 725.703560] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 725.703560] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] self._vmops.spawn(context, instance, image_meta, injected_files, [ 725.703560] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 725.703560] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] vm_ref = self.build_virtual_machine(instance, [ 725.703560] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 725.703560] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] vif_infos = vmwarevif.get_vif_info(self._session, [ 725.703560] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 725.703913] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] for vif in network_info: [ 725.703913] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 725.703913] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] return self._sync_wrapper(fn, *args, **kwargs) [ 725.703913] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 725.703913] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] self.wait() [ 725.703913] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 725.703913] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] self[:] = self._gt.wait() [ 725.703913] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 725.703913] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] return self._exit_event.wait() [ 725.703913] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 725.703913] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] current.throw(*self._exc) [ 725.703913] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.703913] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] result = function(*args, **kwargs) [ 725.704272] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 725.704272] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] return func(*args, **kwargs) [ 725.704272] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.704272] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] raise e [ 725.704272] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.704272] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] nwinfo = self.network_api.allocate_for_instance( [ 725.704272] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.704272] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] created_port_ids = self._update_ports_for_instance( [ 725.704272] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.704272] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] with excutils.save_and_reraise_exception(): [ 725.704272] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.704272] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] self.force_reraise() [ 725.704272] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.704626] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] raise self.value [ 725.704626] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.704626] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] updated_port = self._update_port( [ 725.704626] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.704626] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] _ensure_no_port_binding_failure(port) [ 725.704626] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.704626] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] raise exception.PortBindingFailed(port_id=port['id']) [ 725.704626] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] nova.exception.PortBindingFailed: Binding failed for port da878503-b70a-4e57-b4d1-eacf4bb3c3f3, please check neutron logs for more information. [ 725.704626] env[61970]: ERROR nova.compute.manager [instance: 9d71be92-36ef-4523-919a-931110d73248] [ 725.704626] env[61970]: DEBUG nova.compute.utils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Binding failed for port da878503-b70a-4e57-b4d1-eacf4bb3c3f3, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 725.705549] env[61970]: DEBUG oslo_concurrency.lockutils [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.444s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.706970] env[61970]: INFO nova.compute.claims [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 725.709996] env[61970]: DEBUG nova.compute.manager [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 725.712520] env[61970]: DEBUG nova.compute.manager [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Build of instance 9d71be92-36ef-4523-919a-931110d73248 was re-scheduled: Binding failed for port da878503-b70a-4e57-b4d1-eacf4bb3c3f3, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 725.712940] env[61970]: DEBUG nova.compute.manager [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 725.713357] env[61970]: DEBUG oslo_concurrency.lockutils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Acquiring lock "refresh_cache-9d71be92-36ef-4523-919a-931110d73248" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.713357] env[61970]: DEBUG oslo_concurrency.lockutils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Acquired lock "refresh_cache-9d71be92-36ef-4523-919a-931110d73248" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.713490] env[61970]: DEBUG nova.network.neutron [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 725.744699] env[61970]: DEBUG nova.virt.hardware [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 725.744946] env[61970]: DEBUG nova.virt.hardware [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 725.745116] env[61970]: DEBUG nova.virt.hardware [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 725.745295] env[61970]: DEBUG nova.virt.hardware [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 725.745439] env[61970]: DEBUG nova.virt.hardware [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 725.745581] env[61970]: DEBUG nova.virt.hardware [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 725.745783] env[61970]: DEBUG nova.virt.hardware [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 725.745937] env[61970]: DEBUG nova.virt.hardware [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 725.746305] env[61970]: DEBUG nova.virt.hardware [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 725.746487] env[61970]: DEBUG nova.virt.hardware [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 725.746662] env[61970]: DEBUG nova.virt.hardware [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.747570] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc63192-34bb-4e39-8ced-5405d21d5aa6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.756136] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6443aa2a-9c09-46ef-bb88-5ee20a4ef7b1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.770370] env[61970]: ERROR nova.compute.manager [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 091e50a0-7fd8-4d7c-9035-4a29f32819ae, please check neutron logs for more information. [ 725.770370] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Traceback (most recent call last): [ 725.770370] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 725.770370] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] yield resources [ 725.770370] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 725.770370] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] self.driver.spawn(context, instance, image_meta, [ 725.770370] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 725.770370] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] self._vmops.spawn(context, instance, image_meta, injected_files, [ 725.770370] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 725.770370] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] vm_ref = self.build_virtual_machine(instance, [ 725.770370] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 725.770838] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] vif_infos = vmwarevif.get_vif_info(self._session, [ 725.770838] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 725.770838] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] for vif in network_info: [ 725.770838] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 725.770838] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] return self._sync_wrapper(fn, *args, **kwargs) [ 725.770838] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 725.770838] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] self.wait() [ 725.770838] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 725.770838] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] self[:] = self._gt.wait() [ 725.770838] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 725.770838] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] return self._exit_event.wait() [ 725.770838] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 725.770838] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] current.throw(*self._exc) [ 725.771291] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.771291] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] result = function(*args, **kwargs) [ 725.771291] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 725.771291] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] return func(*args, **kwargs) [ 725.771291] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.771291] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] raise e [ 725.771291] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.771291] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] nwinfo = self.network_api.allocate_for_instance( [ 725.771291] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.771291] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] created_port_ids = self._update_ports_for_instance( [ 725.771291] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.771291] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] with excutils.save_and_reraise_exception(): [ 725.771291] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.771686] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] self.force_reraise() [ 725.771686] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.771686] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] raise self.value [ 725.771686] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.771686] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] updated_port = self._update_port( [ 725.771686] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.771686] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] _ensure_no_port_binding_failure(port) [ 725.771686] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.771686] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] raise exception.PortBindingFailed(port_id=port['id']) [ 725.771686] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] nova.exception.PortBindingFailed: Binding failed for port 091e50a0-7fd8-4d7c-9035-4a29f32819ae, please check neutron logs for more information. [ 725.771686] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] [ 725.771686] env[61970]: INFO nova.compute.manager [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Terminating instance [ 725.772867] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Acquiring lock "refresh_cache-762efd5e-c95e-4718-98de-2e0b05226c06" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.813098] env[61970]: DEBUG nova.virt.hardware [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 725.813341] env[61970]: DEBUG nova.virt.hardware [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 725.813491] env[61970]: DEBUG nova.virt.hardware [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 725.813664] env[61970]: DEBUG nova.virt.hardware [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 725.813808] env[61970]: DEBUG nova.virt.hardware [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 725.813942] env[61970]: DEBUG nova.virt.hardware [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 725.814155] env[61970]: DEBUG nova.virt.hardware [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 725.814315] env[61970]: DEBUG nova.virt.hardware [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 725.814477] env[61970]: DEBUG nova.virt.hardware [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 725.814628] env[61970]: DEBUG nova.virt.hardware [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 725.814853] env[61970]: DEBUG nova.virt.hardware [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.815658] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6db783a-f149-42d9-bfb3-cc27cdc2dc19 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.823382] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc72e6e-4897-45d2-88f5-0479163167e3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.837361] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Instance VIF info [] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 725.842851] env[61970]: DEBUG oslo.service.loopingcall [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.845113] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 725.847242] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d683c6df-c837-4d44-bfe3-13684a62b02d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.864543] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 725.864543] env[61970]: value = "task-1355530" [ 725.864543] env[61970]: _type = "Task" [ 725.864543] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.872722] env[61970]: DEBUG oslo_concurrency.lockutils [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.876142] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355530, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.974116] env[61970]: DEBUG nova.network.neutron [req-6fb19add-3e77-4c61-a624-0bd46d3c632e req-59e507a9-2fdf-460a-b7ef-4053d56443eb service nova] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.083988] env[61970]: DEBUG nova.network.neutron [req-6fb19add-3e77-4c61-a624-0bd46d3c632e req-59e507a9-2fdf-460a-b7ef-4053d56443eb service nova] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.232124] env[61970]: DEBUG nova.network.neutron [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.340048] env[61970]: DEBUG nova.network.neutron [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.373981] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355530, 'name': CreateVM_Task, 'duration_secs': 0.235591} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.374158] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 726.374552] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.374709] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.375052] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 726.375295] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a55fc9ad-7426-4dc8-8400-c12d205adc5d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.379500] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 726.379500] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5203a3cc-5d5c-b674-652b-c9aba80ff196" [ 726.379500] env[61970]: _type = "Task" [ 726.379500] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.388433] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5203a3cc-5d5c-b674-652b-c9aba80ff196, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.587666] env[61970]: DEBUG oslo_concurrency.lockutils [req-6fb19add-3e77-4c61-a624-0bd46d3c632e req-59e507a9-2fdf-460a-b7ef-4053d56443eb service nova] Releasing lock "refresh_cache-762efd5e-c95e-4718-98de-2e0b05226c06" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.588089] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Acquired lock "refresh_cache-762efd5e-c95e-4718-98de-2e0b05226c06" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.588278] env[61970]: DEBUG nova.network.neutron [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 726.842480] env[61970]: DEBUG oslo_concurrency.lockutils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Releasing lock "refresh_cache-9d71be92-36ef-4523-919a-931110d73248" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.842763] env[61970]: DEBUG nova.compute.manager [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 726.843415] env[61970]: DEBUG nova.compute.manager [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 726.843415] env[61970]: DEBUG nova.network.neutron [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 726.865927] env[61970]: DEBUG nova.network.neutron [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.891969] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5203a3cc-5d5c-b674-652b-c9aba80ff196, 'name': SearchDatastore_Task, 'duration_secs': 0.008891} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.892285] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.892554] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 726.892762] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.892910] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.893096] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 726.893355] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-013d9fbd-d258-4d03-ab04-5bfb2f7beb8a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.901311] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 726.902103] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 726.902205] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24f0b2c5-a874-45e9-af9b-40696af4b175 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.907371] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 726.907371] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52172c7c-8ca0-2a22-a105-d5d3d6efb71e" [ 726.907371] env[61970]: _type = "Task" [ 726.907371] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.918707] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52172c7c-8ca0-2a22-a105-d5d3d6efb71e, 'name': SearchDatastore_Task, 'duration_secs': 0.007831} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.919386] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90aa2926-deff-4e56-bec9-a3f4c0f3f076 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.926205] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 726.926205] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a71806-ee94-609e-71eb-b73ddff339e2" [ 726.926205] env[61970]: _type = "Task" [ 726.926205] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.933761] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a71806-ee94-609e-71eb-b73ddff339e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.094288] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030e008e-8bb5-49d3-8a55-69115dfc166a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.102998] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3c83bf4-726c-436a-a5c4-cf5b94297cbb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.136637] env[61970]: DEBUG nova.network.neutron [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.138952] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76866cd-f500-4253-ac77-639adaa6757e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.147731] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81a2f20c-3ab8-4f65-9085-c429441a90fe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.166283] env[61970]: DEBUG nova.compute.provider_tree [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.269020] env[61970]: DEBUG nova.network.neutron [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.374695] env[61970]: DEBUG nova.network.neutron [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.436719] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a71806-ee94-609e-71eb-b73ddff339e2, 'name': SearchDatastore_Task, 'duration_secs': 0.007887} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.436980] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.437244] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 915c3b6c-640b-430f-b264-40bf85c642b9/915c3b6c-640b-430f-b264-40bf85c642b9.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 727.437500] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-efaf6cb2-b29d-43b2-b8f8-7573aa8b7aff {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.443391] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 727.443391] env[61970]: value = "task-1355531" [ 727.443391] env[61970]: _type = "Task" [ 727.443391] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.450538] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355531, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.471354] env[61970]: DEBUG nova.compute.manager [req-cd784708-dc6b-4b43-8b2e-7a59d00f91aa req-0547e178-c9bd-4f9c-b4ac-6e2984f44de3 service nova] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Received event network-vif-deleted-091e50a0-7fd8-4d7c-9035-4a29f32819ae {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.671848] env[61970]: DEBUG nova.scheduler.client.report [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 727.773773] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Releasing lock "refresh_cache-762efd5e-c95e-4718-98de-2e0b05226c06" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.773773] env[61970]: DEBUG nova.compute.manager [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 727.773773] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 727.773773] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5e6b3b7c-1ddf-4268-b548-c0965ea8da93 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.783418] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac2215e-763b-4455-9908-bd8921285b95 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.807035] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 762efd5e-c95e-4718-98de-2e0b05226c06 could not be found. [ 727.807333] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 727.807526] env[61970]: INFO nova.compute.manager [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Took 0.03 seconds to destroy the instance on the hypervisor. [ 727.807849] env[61970]: DEBUG oslo.service.loopingcall [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 727.808100] env[61970]: DEBUG nova.compute.manager [-] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 727.808784] env[61970]: DEBUG nova.network.neutron [-] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 727.831131] env[61970]: DEBUG nova.network.neutron [-] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.876370] env[61970]: INFO nova.compute.manager [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] [instance: 9d71be92-36ef-4523-919a-931110d73248] Took 1.03 seconds to deallocate network for instance. [ 727.953729] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355531, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463835} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.954157] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 915c3b6c-640b-430f-b264-40bf85c642b9/915c3b6c-640b-430f-b264-40bf85c642b9.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 727.954424] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 727.954705] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-659481ff-5dd5-4d77-a559-738893975341 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.960903] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 727.960903] env[61970]: value = "task-1355532" [ 727.960903] env[61970]: _type = "Task" [ 727.960903] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.971610] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355532, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.176638] env[61970]: DEBUG oslo_concurrency.lockutils [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.177682] env[61970]: DEBUG nova.compute.manager [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 728.179769] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.774s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.179954] env[61970]: DEBUG nova.objects.instance [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61970) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 728.333720] env[61970]: DEBUG nova.network.neutron [-] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.470356] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355532, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057571} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.470615] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 728.471444] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7491db89-1ac5-468f-a818-54018bba434a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.492012] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] 915c3b6c-640b-430f-b264-40bf85c642b9/915c3b6c-640b-430f-b264-40bf85c642b9.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 728.492594] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1ca3f3a-d9cb-49bd-829c-f61c19f7afa8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.512240] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 728.512240] env[61970]: value = "task-1355533" [ 728.512240] env[61970]: _type = "Task" [ 728.512240] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.519763] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355533, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.688140] env[61970]: DEBUG nova.compute.utils [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 728.690301] env[61970]: DEBUG nova.compute.manager [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 728.690301] env[61970]: DEBUG nova.network.neutron [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 728.749856] env[61970]: DEBUG nova.policy [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b92bec8029c04f7e96372a2df6406f1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6a554323cee1457c851018b012642cf8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 728.839366] env[61970]: INFO nova.compute.manager [-] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Took 1.03 seconds to deallocate network for instance. [ 728.844871] env[61970]: DEBUG nova.compute.claims [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 728.846249] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.911980] env[61970]: INFO nova.scheduler.client.report [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Deleted allocations for instance 9d71be92-36ef-4523-919a-931110d73248 [ 729.023367] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355533, 'name': ReconfigVM_Task, 'duration_secs': 0.27446} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.023860] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Reconfigured VM instance instance-00000027 to attach disk [datastore2] 915c3b6c-640b-430f-b264-40bf85c642b9/915c3b6c-640b-430f-b264-40bf85c642b9.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 729.028017] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7bcf8933-95ed-49fc-8ed9-9059a196b799 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.031024] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 729.031024] env[61970]: value = "task-1355534" [ 729.031024] env[61970]: _type = "Task" [ 729.031024] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.040022] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355534, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.057920] env[61970]: DEBUG nova.network.neutron [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Successfully created port: 6e1af134-e574-4f35-90e1-a639108500c5 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 729.190880] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae21a5a-710a-4ee2-88c8-b9397749018a tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.192142] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.842s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.199024] env[61970]: DEBUG nova.compute.manager [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 729.423088] env[61970]: DEBUG oslo_concurrency.lockutils [None req-80e4a5dd-7321-42b4-b4d3-0702209b1b11 tempest-ServerGroupTestJSON-1016305505 tempest-ServerGroupTestJSON-1016305505-project-member] Lock "9d71be92-36ef-4523-919a-931110d73248" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.175s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.545265] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355534, 'name': Rename_Task, 'duration_secs': 0.1395} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.545530] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 729.545765] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f1893a72-5519-47ff-af8d-a882f2beaeaf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.552472] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 729.552472] env[61970]: value = "task-1355535" [ 729.552472] env[61970]: _type = "Task" [ 729.552472] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.564691] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355535, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.926390] env[61970]: DEBUG nova.compute.manager [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 729.980118] env[61970]: DEBUG nova.compute.manager [req-f532709b-4eb2-4972-ac23-a0aae4d696d2 req-e1022634-0581-46ab-9721-52fd0175f0bc service nova] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Received event network-changed-6e1af134-e574-4f35-90e1-a639108500c5 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 729.980315] env[61970]: DEBUG nova.compute.manager [req-f532709b-4eb2-4972-ac23-a0aae4d696d2 req-e1022634-0581-46ab-9721-52fd0175f0bc service nova] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Refreshing instance network info cache due to event network-changed-6e1af134-e574-4f35-90e1-a639108500c5. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 729.980534] env[61970]: DEBUG oslo_concurrency.lockutils [req-f532709b-4eb2-4972-ac23-a0aae4d696d2 req-e1022634-0581-46ab-9721-52fd0175f0bc service nova] Acquiring lock "refresh_cache-a540095e-563e-4059-b861-5d0e1b4995f4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.980739] env[61970]: DEBUG oslo_concurrency.lockutils [req-f532709b-4eb2-4972-ac23-a0aae4d696d2 req-e1022634-0581-46ab-9721-52fd0175f0bc service nova] Acquired lock "refresh_cache-a540095e-563e-4059-b861-5d0e1b4995f4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.980814] env[61970]: DEBUG nova.network.neutron [req-f532709b-4eb2-4972-ac23-a0aae4d696d2 req-e1022634-0581-46ab-9721-52fd0175f0bc service nova] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Refreshing network info cache for port 6e1af134-e574-4f35-90e1-a639108500c5 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 730.063167] env[61970]: DEBUG oslo_vmware.api [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355535, 'name': PowerOnVM_Task, 'duration_secs': 0.469064} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.063640] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 730.063872] env[61970]: DEBUG nova.compute.manager [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 730.067667] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9e8a266-637c-40e7-9a4b-8ce599f18563 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.113019] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e70cac-00a0-4f16-b405-620eeab83fad {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.121067] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee7ab6a-91ca-496a-866b-c50d57d5302a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.153524] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dc02b46-6c1e-404f-b5c7-b05c057e0677 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.161169] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-539b7bf8-67d1-4454-82b5-075ac3e8c0f2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.176373] env[61970]: DEBUG nova.compute.provider_tree [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.178877] env[61970]: ERROR nova.compute.manager [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6e1af134-e574-4f35-90e1-a639108500c5, please check neutron logs for more information. [ 730.178877] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 730.178877] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.178877] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 730.178877] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.178877] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 730.178877] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.178877] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 730.178877] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.178877] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 730.178877] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.178877] env[61970]: ERROR nova.compute.manager raise self.value [ 730.178877] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.178877] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 730.178877] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.178877] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 730.179537] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.179537] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 730.179537] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6e1af134-e574-4f35-90e1-a639108500c5, please check neutron logs for more information. [ 730.179537] env[61970]: ERROR nova.compute.manager [ 730.179537] env[61970]: Traceback (most recent call last): [ 730.179537] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 730.179537] env[61970]: listener.cb(fileno) [ 730.179537] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.179537] env[61970]: result = function(*args, **kwargs) [ 730.179537] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 730.179537] env[61970]: return func(*args, **kwargs) [ 730.179537] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 730.179537] env[61970]: raise e [ 730.179537] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.179537] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 730.179537] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.179537] env[61970]: created_port_ids = self._update_ports_for_instance( [ 730.179537] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.179537] env[61970]: with excutils.save_and_reraise_exception(): [ 730.179537] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.179537] env[61970]: self.force_reraise() [ 730.179537] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.179537] env[61970]: raise self.value [ 730.179537] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.179537] env[61970]: updated_port = self._update_port( [ 730.179537] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.179537] env[61970]: _ensure_no_port_binding_failure(port) [ 730.179537] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.179537] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 730.180401] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 6e1af134-e574-4f35-90e1-a639108500c5, please check neutron logs for more information. [ 730.180401] env[61970]: Removing descriptor: 17 [ 730.205577] env[61970]: DEBUG nova.compute.manager [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 730.230955] env[61970]: DEBUG nova.virt.hardware [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 730.231232] env[61970]: DEBUG nova.virt.hardware [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 730.231388] env[61970]: DEBUG nova.virt.hardware [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 730.231568] env[61970]: DEBUG nova.virt.hardware [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 730.231723] env[61970]: DEBUG nova.virt.hardware [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 730.231880] env[61970]: DEBUG nova.virt.hardware [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 730.232101] env[61970]: DEBUG nova.virt.hardware [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 730.232262] env[61970]: DEBUG nova.virt.hardware [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 730.232429] env[61970]: DEBUG nova.virt.hardware [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 730.232605] env[61970]: DEBUG nova.virt.hardware [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 730.232802] env[61970]: DEBUG nova.virt.hardware [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 730.233649] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286ad39f-9c6a-45a5-b8b3-32e23a5e6656 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.243417] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c973dd-ffd0-4e24-800a-a7e0d987cb4b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.255625] env[61970]: ERROR nova.compute.manager [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6e1af134-e574-4f35-90e1-a639108500c5, please check neutron logs for more information. [ 730.255625] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Traceback (most recent call last): [ 730.255625] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 730.255625] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] yield resources [ 730.255625] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 730.255625] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] self.driver.spawn(context, instance, image_meta, [ 730.255625] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 730.255625] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 730.255625] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 730.255625] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] vm_ref = self.build_virtual_machine(instance, [ 730.255625] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 730.256093] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 730.256093] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 730.256093] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] for vif in network_info: [ 730.256093] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 730.256093] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] return self._sync_wrapper(fn, *args, **kwargs) [ 730.256093] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 730.256093] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] self.wait() [ 730.256093] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 730.256093] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] self[:] = self._gt.wait() [ 730.256093] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 730.256093] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] return self._exit_event.wait() [ 730.256093] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 730.256093] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] current.throw(*self._exc) [ 730.256452] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.256452] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] result = function(*args, **kwargs) [ 730.256452] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 730.256452] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] return func(*args, **kwargs) [ 730.256452] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 730.256452] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] raise e [ 730.256452] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.256452] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] nwinfo = self.network_api.allocate_for_instance( [ 730.256452] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.256452] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] created_port_ids = self._update_ports_for_instance( [ 730.256452] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.256452] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] with excutils.save_and_reraise_exception(): [ 730.256452] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.256798] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] self.force_reraise() [ 730.256798] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.256798] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] raise self.value [ 730.256798] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.256798] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] updated_port = self._update_port( [ 730.256798] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.256798] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] _ensure_no_port_binding_failure(port) [ 730.256798] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.256798] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] raise exception.PortBindingFailed(port_id=port['id']) [ 730.256798] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] nova.exception.PortBindingFailed: Binding failed for port 6e1af134-e574-4f35-90e1-a639108500c5, please check neutron logs for more information. [ 730.256798] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] [ 730.256798] env[61970]: INFO nova.compute.manager [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Terminating instance [ 730.257744] env[61970]: DEBUG oslo_concurrency.lockutils [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Acquiring lock "refresh_cache-a540095e-563e-4059-b861-5d0e1b4995f4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.451440] env[61970]: DEBUG oslo_concurrency.lockutils [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.499763] env[61970]: DEBUG nova.network.neutron [req-f532709b-4eb2-4972-ac23-a0aae4d696d2 req-e1022634-0581-46ab-9721-52fd0175f0bc service nova] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.583104] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.609512] env[61970]: DEBUG nova.network.neutron [req-f532709b-4eb2-4972-ac23-a0aae4d696d2 req-e1022634-0581-46ab-9721-52fd0175f0bc service nova] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.680290] env[61970]: DEBUG nova.scheduler.client.report [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 731.037657] env[61970]: INFO nova.compute.manager [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Rebuilding instance [ 731.095752] env[61970]: DEBUG nova.compute.manager [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 731.096630] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3b49e4-5f41-44a2-bdf7-e66659642b88 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.112449] env[61970]: DEBUG oslo_concurrency.lockutils [req-f532709b-4eb2-4972-ac23-a0aae4d696d2 req-e1022634-0581-46ab-9721-52fd0175f0bc service nova] Releasing lock "refresh_cache-a540095e-563e-4059-b861-5d0e1b4995f4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.112835] env[61970]: DEBUG oslo_concurrency.lockutils [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Acquired lock "refresh_cache-a540095e-563e-4059-b861-5d0e1b4995f4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.113061] env[61970]: DEBUG nova.network.neutron [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 731.185293] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.993s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.185892] env[61970]: ERROR nova.compute.manager [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b758644c-56b6-4a94-b6e6-0d038fd951b4, please check neutron logs for more information. [ 731.185892] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Traceback (most recent call last): [ 731.185892] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 731.185892] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] self.driver.spawn(context, instance, image_meta, [ 731.185892] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 731.185892] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 731.185892] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 731.185892] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] vm_ref = self.build_virtual_machine(instance, [ 731.185892] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 731.185892] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] vif_infos = vmwarevif.get_vif_info(self._session, [ 731.185892] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 731.186279] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] for vif in network_info: [ 731.186279] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 731.186279] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] return self._sync_wrapper(fn, *args, **kwargs) [ 731.186279] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 731.186279] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] self.wait() [ 731.186279] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 731.186279] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] self[:] = self._gt.wait() [ 731.186279] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 731.186279] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] return self._exit_event.wait() [ 731.186279] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 731.186279] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] current.throw(*self._exc) [ 731.186279] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.186279] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] result = function(*args, **kwargs) [ 731.186679] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 731.186679] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] return func(*args, **kwargs) [ 731.186679] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 731.186679] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] raise e [ 731.186679] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.186679] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] nwinfo = self.network_api.allocate_for_instance( [ 731.186679] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 731.186679] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] created_port_ids = self._update_ports_for_instance( [ 731.186679] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 731.186679] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] with excutils.save_and_reraise_exception(): [ 731.186679] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.186679] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] self.force_reraise() [ 731.186679] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.187096] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] raise self.value [ 731.187096] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 731.187096] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] updated_port = self._update_port( [ 731.187096] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.187096] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] _ensure_no_port_binding_failure(port) [ 731.187096] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.187096] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] raise exception.PortBindingFailed(port_id=port['id']) [ 731.187096] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] nova.exception.PortBindingFailed: Binding failed for port b758644c-56b6-4a94-b6e6-0d038fd951b4, please check neutron logs for more information. [ 731.187096] env[61970]: ERROR nova.compute.manager [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] [ 731.187096] env[61970]: DEBUG nova.compute.utils [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Binding failed for port b758644c-56b6-4a94-b6e6-0d038fd951b4, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 731.187838] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.351s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.189328] env[61970]: INFO nova.compute.claims [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.192212] env[61970]: DEBUG nova.compute.manager [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Build of instance 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9 was re-scheduled: Binding failed for port b758644c-56b6-4a94-b6e6-0d038fd951b4, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 731.192454] env[61970]: DEBUG nova.compute.manager [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 731.192677] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Acquiring lock "refresh_cache-630145f3-cd01-41fe-b9c5-7ebf44b1e6d9" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.192821] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Acquired lock "refresh_cache-630145f3-cd01-41fe-b9c5-7ebf44b1e6d9" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.192976] env[61970]: DEBUG nova.network.neutron [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 731.609094] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 731.609435] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67499066-f991-4305-81f7-057f55d65a1c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.622253] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Waiting for the task: (returnval){ [ 731.622253] env[61970]: value = "task-1355536" [ 731.622253] env[61970]: _type = "Task" [ 731.622253] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.637096] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': task-1355536, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.639374] env[61970]: DEBUG nova.network.neutron [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.728907] env[61970]: DEBUG nova.network.neutron [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.771749] env[61970]: DEBUG nova.network.neutron [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.831459] env[61970]: DEBUG nova.network.neutron [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.029347] env[61970]: DEBUG nova.compute.manager [req-932217e7-43f7-4fa6-91d4-6efec61e1f6c req-ebc49ed0-d38e-47e4-8099-37a7a34b8aef service nova] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Received event network-vif-deleted-6e1af134-e574-4f35-90e1-a639108500c5 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 732.136543] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': task-1355536, 'name': PowerOffVM_Task, 'duration_secs': 0.110025} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.136975] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 732.137252] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 732.138285] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4945c403-f7da-4c2d-8669-dd8b71a2a1bd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.146143] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 732.146640] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6fedc2d4-6a1c-4fbd-8b2c-cc0e77b70c39 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.170365] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 732.170365] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 732.170598] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Deleting the datastore file [datastore2] 915c3b6c-640b-430f-b264-40bf85c642b9 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 732.170803] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-23b25e9a-49f5-4eec-bd55-1e8432dca7f7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.179087] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Waiting for the task: (returnval){ [ 732.179087] env[61970]: value = "task-1355538" [ 732.179087] env[61970]: _type = "Task" [ 732.179087] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.190029] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': task-1355538, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.275590] env[61970]: DEBUG oslo_concurrency.lockutils [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Releasing lock "refresh_cache-a540095e-563e-4059-b861-5d0e1b4995f4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.276321] env[61970]: DEBUG nova.compute.manager [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 732.276413] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 732.276952] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf9eaf17-12ec-4449-a261-79b295b90374 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.285119] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50fe517-1a59-4c1b-b1e0-1532a8c5a2c4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.309982] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a540095e-563e-4059-b861-5d0e1b4995f4 could not be found. [ 732.310317] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 732.310521] env[61970]: INFO nova.compute.manager [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 732.310826] env[61970]: DEBUG oslo.service.loopingcall [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 732.313525] env[61970]: DEBUG nova.compute.manager [-] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 732.313639] env[61970]: DEBUG nova.network.neutron [-] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 732.331291] env[61970]: DEBUG nova.network.neutron [-] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.333792] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Releasing lock "refresh_cache-630145f3-cd01-41fe-b9c5-7ebf44b1e6d9" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.334061] env[61970]: DEBUG nova.compute.manager [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 732.334298] env[61970]: DEBUG nova.compute.manager [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 732.334495] env[61970]: DEBUG nova.network.neutron [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 732.348463] env[61970]: DEBUG nova.network.neutron [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.575715] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04116109-f1be-46fb-86c3-1515d681d8d1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.583400] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16357cd6-2e8b-4a91-9169-ddf38b66de29 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.613754] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2752b3a5-6eb5-4afc-b54d-342e927f3f5c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.621906] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4285097-8609-46c1-9ef2-7ad8d7654829 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.636446] env[61970]: DEBUG nova.compute.provider_tree [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.688324] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': task-1355538, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101698} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.688621] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 732.688834] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 732.689023] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 732.833884] env[61970]: DEBUG nova.network.neutron [-] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.853515] env[61970]: DEBUG nova.network.neutron [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.139469] env[61970]: DEBUG nova.scheduler.client.report [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 733.337676] env[61970]: INFO nova.compute.manager [-] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Took 1.02 seconds to deallocate network for instance. [ 733.339879] env[61970]: DEBUG nova.compute.claims [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 733.340068] env[61970]: DEBUG oslo_concurrency.lockutils [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.355982] env[61970]: INFO nova.compute.manager [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] [instance: 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9] Took 1.02 seconds to deallocate network for instance. [ 733.645826] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.457s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.645826] env[61970]: DEBUG nova.compute.manager [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 733.648831] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.142s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.648831] env[61970]: DEBUG nova.objects.instance [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Lazy-loading 'resources' on Instance uuid 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 733.719495] env[61970]: DEBUG nova.virt.hardware [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 733.719780] env[61970]: DEBUG nova.virt.hardware [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 733.719934] env[61970]: DEBUG nova.virt.hardware [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 733.720124] env[61970]: DEBUG nova.virt.hardware [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 733.720271] env[61970]: DEBUG nova.virt.hardware [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 733.720416] env[61970]: DEBUG nova.virt.hardware [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 733.720618] env[61970]: DEBUG nova.virt.hardware [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 733.720776] env[61970]: DEBUG nova.virt.hardware [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 733.720941] env[61970]: DEBUG nova.virt.hardware [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 733.721116] env[61970]: DEBUG nova.virt.hardware [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 733.721292] env[61970]: DEBUG nova.virt.hardware [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 733.722158] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d10046-818e-4eab-8f07-c1026478e86e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.730934] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3951bb03-9af5-4eee-9b1d-328a9373093d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.745838] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Instance VIF info [] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 733.751306] env[61970]: DEBUG oslo.service.loopingcall [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 733.751543] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 733.751799] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bd21fe26-639b-4492-91e7-abfdad58acbc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.768217] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 733.768217] env[61970]: value = "task-1355539" [ 733.768217] env[61970]: _type = "Task" [ 733.768217] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.775632] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355539, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.154073] env[61970]: DEBUG nova.compute.utils [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 734.155785] env[61970]: DEBUG nova.compute.manager [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 734.156136] env[61970]: DEBUG nova.network.neutron [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 734.216823] env[61970]: DEBUG nova.policy [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64082edf5f2e4621b5f3586e1792e42a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9235eeeb6fa4ccf959ffbea456f3694', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 734.277534] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355539, 'name': CreateVM_Task, 'duration_secs': 0.257183} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.279876] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 734.280711] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.280711] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.282027] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 734.282271] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06ebd2e4-108c-447b-b257-27e77fba01f7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.286760] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Waiting for the task: (returnval){ [ 734.286760] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52971f87-7852-4b1c-8d57-bb1390f2e917" [ 734.286760] env[61970]: _type = "Task" [ 734.286760] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.297149] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52971f87-7852-4b1c-8d57-bb1390f2e917, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.389020] env[61970]: INFO nova.scheduler.client.report [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Deleted allocations for instance 630145f3-cd01-41fe-b9c5-7ebf44b1e6d9 [ 734.517028] env[61970]: DEBUG nova.network.neutron [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Successfully created port: dab2accb-1541-4bcf-8bde-d36049571418 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 734.531194] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-146e7d2a-1795-4636-81ca-d020990569f6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.538758] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e507e7-e435-426d-ae73-ad6b5c21468c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.568209] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b2712ae-19bf-4f50-9a75-ec9109b1dc6c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.574989] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9cac17c-52b3-41db-bd95-a759be955ef8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.587743] env[61970]: DEBUG nova.compute.provider_tree [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.659517] env[61970]: DEBUG nova.compute.manager [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 734.796960] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52971f87-7852-4b1c-8d57-bb1390f2e917, 'name': SearchDatastore_Task, 'duration_secs': 0.00881} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.797279] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.797512] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 734.797743] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.797887] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.798078] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 734.798322] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7308285e-f19f-49b7-a504-4ecf35bf9004 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.806293] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 734.806474] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 734.807389] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3067987-c283-41f9-8872-39cfb3783a15 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.812358] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Waiting for the task: (returnval){ [ 734.812358] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f58761-e1f1-cf21-cb61-a49fd43a765d" [ 734.812358] env[61970]: _type = "Task" [ 734.812358] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.819933] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f58761-e1f1-cf21-cb61-a49fd43a765d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.897047] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b943ecc9-ba13-4ca8-88b8-3fe933e34c3f tempest-FloatingIPsAssociationNegativeTestJSON-531860821 tempest-FloatingIPsAssociationNegativeTestJSON-531860821-project-member] Lock "630145f3-cd01-41fe-b9c5-7ebf44b1e6d9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.770s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.093018] env[61970]: DEBUG nova.scheduler.client.report [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 735.302105] env[61970]: DEBUG nova.compute.manager [req-d4e46ced-e879-41c7-9caf-cf0ccdbe60cd req-d1394115-1cda-4f65-adb3-9f2cbb07a23f service nova] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Received event network-changed-dab2accb-1541-4bcf-8bde-d36049571418 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 735.303100] env[61970]: DEBUG nova.compute.manager [req-d4e46ced-e879-41c7-9caf-cf0ccdbe60cd req-d1394115-1cda-4f65-adb3-9f2cbb07a23f service nova] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Refreshing instance network info cache due to event network-changed-dab2accb-1541-4bcf-8bde-d36049571418. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 735.303100] env[61970]: DEBUG oslo_concurrency.lockutils [req-d4e46ced-e879-41c7-9caf-cf0ccdbe60cd req-d1394115-1cda-4f65-adb3-9f2cbb07a23f service nova] Acquiring lock "refresh_cache-29ac968a-37db-4c0d-9be9-9f577b53c533" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.303100] env[61970]: DEBUG oslo_concurrency.lockutils [req-d4e46ced-e879-41c7-9caf-cf0ccdbe60cd req-d1394115-1cda-4f65-adb3-9f2cbb07a23f service nova] Acquired lock "refresh_cache-29ac968a-37db-4c0d-9be9-9f577b53c533" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.303100] env[61970]: DEBUG nova.network.neutron [req-d4e46ced-e879-41c7-9caf-cf0ccdbe60cd req-d1394115-1cda-4f65-adb3-9f2cbb07a23f service nova] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Refreshing network info cache for port dab2accb-1541-4bcf-8bde-d36049571418 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 735.323685] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f58761-e1f1-cf21-cb61-a49fd43a765d, 'name': SearchDatastore_Task, 'duration_secs': 0.007969} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.324575] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b700f871-06ae-43b3-8ec8-9ccf21dda08b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.329674] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Waiting for the task: (returnval){ [ 735.329674] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525ad519-c6fc-756d-fe4e-c9e309e506ac" [ 735.329674] env[61970]: _type = "Task" [ 735.329674] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.337308] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525ad519-c6fc-756d-fe4e-c9e309e506ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.400608] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 735.568454] env[61970]: ERROR nova.compute.manager [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dab2accb-1541-4bcf-8bde-d36049571418, please check neutron logs for more information. [ 735.568454] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 735.568454] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.568454] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 735.568454] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.568454] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 735.568454] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.568454] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 735.568454] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.568454] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 735.568454] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.568454] env[61970]: ERROR nova.compute.manager raise self.value [ 735.568454] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.568454] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 735.568454] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.568454] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 735.569071] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.569071] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 735.569071] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dab2accb-1541-4bcf-8bde-d36049571418, please check neutron logs for more information. [ 735.569071] env[61970]: ERROR nova.compute.manager [ 735.569071] env[61970]: Traceback (most recent call last): [ 735.569071] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 735.569071] env[61970]: listener.cb(fileno) [ 735.569071] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.569071] env[61970]: result = function(*args, **kwargs) [ 735.569071] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 735.569071] env[61970]: return func(*args, **kwargs) [ 735.569071] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.569071] env[61970]: raise e [ 735.569071] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.569071] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 735.569071] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.569071] env[61970]: created_port_ids = self._update_ports_for_instance( [ 735.569071] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.569071] env[61970]: with excutils.save_and_reraise_exception(): [ 735.569071] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.569071] env[61970]: self.force_reraise() [ 735.569071] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.569071] env[61970]: raise self.value [ 735.569071] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.569071] env[61970]: updated_port = self._update_port( [ 735.569071] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.569071] env[61970]: _ensure_no_port_binding_failure(port) [ 735.569071] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.569071] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 735.570123] env[61970]: nova.exception.PortBindingFailed: Binding failed for port dab2accb-1541-4bcf-8bde-d36049571418, please check neutron logs for more information. [ 735.570123] env[61970]: Removing descriptor: 15 [ 735.597956] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.950s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.600481] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.762s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.602020] env[61970]: INFO nova.compute.claims [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.634278] env[61970]: INFO nova.scheduler.client.report [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Deleted allocations for instance 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8 [ 735.670134] env[61970]: DEBUG nova.compute.manager [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 735.698281] env[61970]: DEBUG nova.virt.hardware [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 735.698569] env[61970]: DEBUG nova.virt.hardware [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 735.702021] env[61970]: DEBUG nova.virt.hardware [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.702021] env[61970]: DEBUG nova.virt.hardware [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 735.702021] env[61970]: DEBUG nova.virt.hardware [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.702021] env[61970]: DEBUG nova.virt.hardware [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 735.702021] env[61970]: DEBUG nova.virt.hardware [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 735.702021] env[61970]: DEBUG nova.virt.hardware [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 735.702279] env[61970]: DEBUG nova.virt.hardware [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 735.702279] env[61970]: DEBUG nova.virt.hardware [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 735.702279] env[61970]: DEBUG nova.virt.hardware [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.702279] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77793580-df2c-420b-b907-2d56db95da8a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.709293] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a946db-aa31-489a-8d83-979e0ba0fe95 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.725856] env[61970]: ERROR nova.compute.manager [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dab2accb-1541-4bcf-8bde-d36049571418, please check neutron logs for more information. [ 735.725856] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Traceback (most recent call last): [ 735.725856] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 735.725856] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] yield resources [ 735.725856] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 735.725856] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] self.driver.spawn(context, instance, image_meta, [ 735.725856] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 735.725856] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] self._vmops.spawn(context, instance, image_meta, injected_files, [ 735.725856] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 735.725856] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] vm_ref = self.build_virtual_machine(instance, [ 735.725856] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 735.726292] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] vif_infos = vmwarevif.get_vif_info(self._session, [ 735.726292] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 735.726292] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] for vif in network_info: [ 735.726292] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 735.726292] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] return self._sync_wrapper(fn, *args, **kwargs) [ 735.726292] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 735.726292] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] self.wait() [ 735.726292] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 735.726292] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] self[:] = self._gt.wait() [ 735.726292] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 735.726292] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] return self._exit_event.wait() [ 735.726292] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 735.726292] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] current.throw(*self._exc) [ 735.726684] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.726684] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] result = function(*args, **kwargs) [ 735.726684] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 735.726684] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] return func(*args, **kwargs) [ 735.726684] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.726684] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] raise e [ 735.726684] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.726684] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] nwinfo = self.network_api.allocate_for_instance( [ 735.726684] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.726684] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] created_port_ids = self._update_ports_for_instance( [ 735.726684] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.726684] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] with excutils.save_and_reraise_exception(): [ 735.726684] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.727099] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] self.force_reraise() [ 735.727099] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.727099] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] raise self.value [ 735.727099] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.727099] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] updated_port = self._update_port( [ 735.727099] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.727099] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] _ensure_no_port_binding_failure(port) [ 735.727099] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.727099] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] raise exception.PortBindingFailed(port_id=port['id']) [ 735.727099] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] nova.exception.PortBindingFailed: Binding failed for port dab2accb-1541-4bcf-8bde-d36049571418, please check neutron logs for more information. [ 735.727099] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] [ 735.727099] env[61970]: INFO nova.compute.manager [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Terminating instance [ 735.728967] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "refresh_cache-29ac968a-37db-4c0d-9be9-9f577b53c533" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.825420] env[61970]: DEBUG nova.network.neutron [req-d4e46ced-e879-41c7-9caf-cf0ccdbe60cd req-d1394115-1cda-4f65-adb3-9f2cbb07a23f service nova] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.839820] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525ad519-c6fc-756d-fe4e-c9e309e506ac, 'name': SearchDatastore_Task, 'duration_secs': 0.009119} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.841194] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.841451] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 915c3b6c-640b-430f-b264-40bf85c642b9/915c3b6c-640b-430f-b264-40bf85c642b9.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 735.841691] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1ccdfe76-02d7-4c41-a80a-4cb5e8784d1b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.849500] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Waiting for the task: (returnval){ [ 735.849500] env[61970]: value = "task-1355540" [ 735.849500] env[61970]: _type = "Task" [ 735.849500] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.858600] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': task-1355540, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.927200] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.951737] env[61970]: DEBUG nova.network.neutron [req-d4e46ced-e879-41c7-9caf-cf0ccdbe60cd req-d1394115-1cda-4f65-adb3-9f2cbb07a23f service nova] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.143377] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1fda2691-3365-4047-9552-ce4a953f13a6 tempest-ServerShowV254Test-334586436 tempest-ServerShowV254Test-334586436-project-member] Lock "9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.314s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.358864] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': task-1355540, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485358} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.359240] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 915c3b6c-640b-430f-b264-40bf85c642b9/915c3b6c-640b-430f-b264-40bf85c642b9.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 736.359497] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 736.359774] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d4bb0166-565d-40dd-9e62-ad02cb3c4ea0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.369195] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Waiting for the task: (returnval){ [ 736.369195] env[61970]: value = "task-1355541" [ 736.369195] env[61970]: _type = "Task" [ 736.369195] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.375914] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': task-1355541, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.456807] env[61970]: DEBUG oslo_concurrency.lockutils [req-d4e46ced-e879-41c7-9caf-cf0ccdbe60cd req-d1394115-1cda-4f65-adb3-9f2cbb07a23f service nova] Releasing lock "refresh_cache-29ac968a-37db-4c0d-9be9-9f577b53c533" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.456807] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "refresh_cache-29ac968a-37db-4c0d-9be9-9f577b53c533" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.456807] env[61970]: DEBUG nova.network.neutron [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 736.886645] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': task-1355541, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069812} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.886645] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 736.886645] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e913ed-d861-4a23-a8e4-3cf12627c7a1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.909947] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] 915c3b6c-640b-430f-b264-40bf85c642b9/915c3b6c-640b-430f-b264-40bf85c642b9.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 736.912816] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e77fe1a8-0e2b-420f-8769-9be99ab4203e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.937839] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Waiting for the task: (returnval){ [ 736.937839] env[61970]: value = "task-1355542" [ 736.937839] env[61970]: _type = "Task" [ 736.937839] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.949729] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': task-1355542, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.986310] env[61970]: DEBUG nova.network.neutron [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.065911] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45179964-c0be-4897-9ee3-8a4d7e4baf7c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.073206] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f9a6c8-4521-4b18-abaf-603ae301d939 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.104281] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c05c8fa-dbb5-4466-aea7-6b8d6ca10ea9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.112281] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3710a2c4-e444-44ba-9cd2-942d9c6d8d58 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.126719] env[61970]: DEBUG nova.compute.provider_tree [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.137629] env[61970]: DEBUG nova.network.neutron [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.331960] env[61970]: DEBUG nova.compute.manager [req-870ac02b-4076-4aa0-acd4-702e0ec35699 req-a7079478-d670-4ff4-bc6d-bb23d32a0ddc service nova] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Received event network-vif-deleted-dab2accb-1541-4bcf-8bde-d36049571418 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 737.446954] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': task-1355542, 'name': ReconfigVM_Task, 'duration_secs': 0.301683} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.447244] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Reconfigured VM instance instance-00000027 to attach disk [datastore2] 915c3b6c-640b-430f-b264-40bf85c642b9/915c3b6c-640b-430f-b264-40bf85c642b9.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 737.447868] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d86fe0a-2ee1-4c4c-831a-e6b200d73e28 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.454219] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Waiting for the task: (returnval){ [ 737.454219] env[61970]: value = "task-1355543" [ 737.454219] env[61970]: _type = "Task" [ 737.454219] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.462017] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': task-1355543, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.635124] env[61970]: DEBUG nova.scheduler.client.report [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.640471] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "refresh_cache-29ac968a-37db-4c0d-9be9-9f577b53c533" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.640556] env[61970]: DEBUG nova.compute.manager [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 737.640935] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 737.641058] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3b44c2d8-70f0-4cf1-8591-2b52067c4a18 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.650318] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6af9d3f-fb6e-4745-967c-bbc9ca3dc445 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.674201] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 737.675275] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 29ac968a-37db-4c0d-9be9-9f577b53c533 could not be found. [ 737.675275] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 737.675414] env[61970]: INFO nova.compute.manager [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Took 0.03 seconds to destroy the instance on the hypervisor. [ 737.675550] env[61970]: DEBUG oslo.service.loopingcall [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 737.675716] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 737.676842] env[61970]: DEBUG nova.compute.manager [-] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 737.676936] env[61970]: DEBUG nova.network.neutron [-] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 737.692195] env[61970]: DEBUG nova.network.neutron [-] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.965748] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': task-1355543, 'name': Rename_Task, 'duration_secs': 0.134527} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.966019] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 737.966276] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a7b6500a-12c0-44e9-93e6-f30273b1ff90 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.972498] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Waiting for the task: (returnval){ [ 737.972498] env[61970]: value = "task-1355544" [ 737.972498] env[61970]: _type = "Task" [ 737.972498] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.979845] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': task-1355544, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.140477] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.540s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.141048] env[61970]: DEBUG nova.compute.manager [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 738.144338] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.789s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.147041] env[61970]: INFO nova.compute.claims [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 738.184188] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.185250] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Starting heal instance info cache {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 738.185250] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Rebuilding the list of instances to heal {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 738.197791] env[61970]: DEBUG nova.network.neutron [-] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.481749] env[61970]: DEBUG oslo_vmware.api [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Task: {'id': task-1355544, 'name': PowerOnVM_Task, 'duration_secs': 0.442964} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.483519] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 738.483519] env[61970]: DEBUG nova.compute.manager [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 738.483519] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d95c546f-17c9-4210-818d-ab3d1347967a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.652452] env[61970]: DEBUG nova.compute.utils [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 738.664237] env[61970]: DEBUG nova.compute.manager [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 738.664416] env[61970]: DEBUG nova.network.neutron [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 738.693853] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 738.694023] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 738.694189] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 738.694270] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 738.694406] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 738.694470] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 738.696345] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "refresh_cache-915c3b6c-640b-430f-b264-40bf85c642b9" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.696345] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquired lock "refresh_cache-915c3b6c-640b-430f-b264-40bf85c642b9" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.696345] env[61970]: DEBUG nova.network.neutron [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Forcefully refreshing network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 738.696345] env[61970]: DEBUG nova.objects.instance [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lazy-loading 'info_cache' on Instance uuid 915c3b6c-640b-430f-b264-40bf85c642b9 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 738.699852] env[61970]: INFO nova.compute.manager [-] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Took 1.02 seconds to deallocate network for instance. [ 738.701742] env[61970]: DEBUG nova.compute.claims [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 738.701930] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.740750] env[61970]: DEBUG nova.policy [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '993dac97555f4703a04c3953fdece450', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f8970fae4b2a48129f3b89430166a056', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 739.000227] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.146897] env[61970]: DEBUG nova.network.neutron [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Successfully created port: 663f1e4e-7249-473d-8538-7f4e66ba7d9c {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 739.165690] env[61970]: DEBUG nova.compute.manager [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 739.623053] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eabdb4ea-18f0-4cfc-bb9e-7d5eea714645 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.630743] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f252bb72-a5ca-4d07-9650-90c9036465a1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.667360] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-900e1aa6-000c-424b-8c19-453f94f43ed3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.679881] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f61d79-5b10-4471-8248-76f5c9865cad {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.696277] env[61970]: DEBUG nova.compute.provider_tree [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.722986] env[61970]: DEBUG nova.network.neutron [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.023285] env[61970]: DEBUG nova.compute.manager [req-b68911a8-73ea-442e-ba6b-405bdc220852 req-d4552b62-7b8e-40d3-be90-8405f3a29cab service nova] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Received event network-changed-663f1e4e-7249-473d-8538-7f4e66ba7d9c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.023445] env[61970]: DEBUG nova.compute.manager [req-b68911a8-73ea-442e-ba6b-405bdc220852 req-d4552b62-7b8e-40d3-be90-8405f3a29cab service nova] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Refreshing instance network info cache due to event network-changed-663f1e4e-7249-473d-8538-7f4e66ba7d9c. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 740.023620] env[61970]: DEBUG oslo_concurrency.lockutils [req-b68911a8-73ea-442e-ba6b-405bdc220852 req-d4552b62-7b8e-40d3-be90-8405f3a29cab service nova] Acquiring lock "refresh_cache-34719fba-dc16-4b43-ab48-517982ca85a4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.023735] env[61970]: DEBUG oslo_concurrency.lockutils [req-b68911a8-73ea-442e-ba6b-405bdc220852 req-d4552b62-7b8e-40d3-be90-8405f3a29cab service nova] Acquired lock "refresh_cache-34719fba-dc16-4b43-ab48-517982ca85a4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.023907] env[61970]: DEBUG nova.network.neutron [req-b68911a8-73ea-442e-ba6b-405bdc220852 req-d4552b62-7b8e-40d3-be90-8405f3a29cab service nova] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Refreshing network info cache for port 663f1e4e-7249-473d-8538-7f4e66ba7d9c {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 740.048029] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Acquiring lock "915c3b6c-640b-430f-b264-40bf85c642b9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.048172] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Lock "915c3b6c-640b-430f-b264-40bf85c642b9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.050033] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Acquiring lock "915c3b6c-640b-430f-b264-40bf85c642b9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.050292] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Lock "915c3b6c-640b-430f-b264-40bf85c642b9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.050421] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Lock "915c3b6c-640b-430f-b264-40bf85c642b9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.052906] env[61970]: INFO nova.compute.manager [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Terminating instance [ 740.054947] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Acquiring lock "refresh_cache-915c3b6c-640b-430f-b264-40bf85c642b9" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.175903] env[61970]: ERROR nova.compute.manager [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 663f1e4e-7249-473d-8538-7f4e66ba7d9c, please check neutron logs for more information. [ 740.175903] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 740.175903] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.175903] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 740.175903] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.175903] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 740.175903] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.175903] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 740.175903] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.175903] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 740.175903] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.175903] env[61970]: ERROR nova.compute.manager raise self.value [ 740.175903] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.175903] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 740.175903] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.175903] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 740.176520] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.176520] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 740.176520] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 663f1e4e-7249-473d-8538-7f4e66ba7d9c, please check neutron logs for more information. [ 740.176520] env[61970]: ERROR nova.compute.manager [ 740.176520] env[61970]: Traceback (most recent call last): [ 740.176520] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 740.176520] env[61970]: listener.cb(fileno) [ 740.176520] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.176520] env[61970]: result = function(*args, **kwargs) [ 740.176520] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 740.176520] env[61970]: return func(*args, **kwargs) [ 740.176520] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.176520] env[61970]: raise e [ 740.176520] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.176520] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 740.176520] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.176520] env[61970]: created_port_ids = self._update_ports_for_instance( [ 740.176520] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.176520] env[61970]: with excutils.save_and_reraise_exception(): [ 740.176520] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.176520] env[61970]: self.force_reraise() [ 740.176520] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.176520] env[61970]: raise self.value [ 740.176520] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.176520] env[61970]: updated_port = self._update_port( [ 740.176520] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.176520] env[61970]: _ensure_no_port_binding_failure(port) [ 740.176520] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.176520] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 740.177445] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 663f1e4e-7249-473d-8538-7f4e66ba7d9c, please check neutron logs for more information. [ 740.177445] env[61970]: Removing descriptor: 15 [ 740.178640] env[61970]: DEBUG nova.compute.manager [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 740.199431] env[61970]: DEBUG nova.scheduler.client.report [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 740.210297] env[61970]: DEBUG nova.virt.hardware [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 740.210297] env[61970]: DEBUG nova.virt.hardware [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 740.210606] env[61970]: DEBUG nova.virt.hardware [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 740.210606] env[61970]: DEBUG nova.virt.hardware [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 740.210756] env[61970]: DEBUG nova.virt.hardware [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 740.210984] env[61970]: DEBUG nova.virt.hardware [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 740.211213] env[61970]: DEBUG nova.virt.hardware [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 740.211373] env[61970]: DEBUG nova.virt.hardware [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 740.211536] env[61970]: DEBUG nova.virt.hardware [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 740.211695] env[61970]: DEBUG nova.virt.hardware [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 740.211890] env[61970]: DEBUG nova.virt.hardware [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 740.212768] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb1a057-3926-4b4d-98e1-36b208230868 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.221752] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d5c72d-fc59-4f6c-a66f-5d40e4c0971e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.237171] env[61970]: ERROR nova.compute.manager [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 663f1e4e-7249-473d-8538-7f4e66ba7d9c, please check neutron logs for more information. [ 740.237171] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Traceback (most recent call last): [ 740.237171] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 740.237171] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] yield resources [ 740.237171] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 740.237171] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] self.driver.spawn(context, instance, image_meta, [ 740.237171] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 740.237171] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 740.237171] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 740.237171] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] vm_ref = self.build_virtual_machine(instance, [ 740.237171] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 740.237890] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] vif_infos = vmwarevif.get_vif_info(self._session, [ 740.237890] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 740.237890] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] for vif in network_info: [ 740.237890] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 740.237890] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] return self._sync_wrapper(fn, *args, **kwargs) [ 740.237890] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 740.237890] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] self.wait() [ 740.237890] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 740.237890] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] self[:] = self._gt.wait() [ 740.237890] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 740.237890] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] return self._exit_event.wait() [ 740.237890] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 740.237890] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] current.throw(*self._exc) [ 740.238732] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.238732] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] result = function(*args, **kwargs) [ 740.238732] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 740.238732] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] return func(*args, **kwargs) [ 740.238732] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.238732] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] raise e [ 740.238732] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.238732] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] nwinfo = self.network_api.allocate_for_instance( [ 740.238732] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.238732] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] created_port_ids = self._update_ports_for_instance( [ 740.238732] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.238732] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] with excutils.save_and_reraise_exception(): [ 740.238732] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.239211] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] self.force_reraise() [ 740.239211] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.239211] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] raise self.value [ 740.239211] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.239211] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] updated_port = self._update_port( [ 740.239211] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.239211] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] _ensure_no_port_binding_failure(port) [ 740.239211] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.239211] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] raise exception.PortBindingFailed(port_id=port['id']) [ 740.239211] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] nova.exception.PortBindingFailed: Binding failed for port 663f1e4e-7249-473d-8538-7f4e66ba7d9c, please check neutron logs for more information. [ 740.239211] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] [ 740.239211] env[61970]: INFO nova.compute.manager [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Terminating instance [ 740.239554] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Acquiring lock "refresh_cache-34719fba-dc16-4b43-ab48-517982ca85a4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.290321] env[61970]: DEBUG nova.network.neutron [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.553185] env[61970]: DEBUG nova.network.neutron [req-b68911a8-73ea-442e-ba6b-405bdc220852 req-d4552b62-7b8e-40d3-be90-8405f3a29cab service nova] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.637779] env[61970]: DEBUG nova.network.neutron [req-b68911a8-73ea-442e-ba6b-405bdc220852 req-d4552b62-7b8e-40d3-be90-8405f3a29cab service nova] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.705067] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.560s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.706861] env[61970]: DEBUG nova.compute.manager [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 740.710572] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.705s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.712218] env[61970]: INFO nova.compute.claims [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 740.795019] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Releasing lock "refresh_cache-915c3b6c-640b-430f-b264-40bf85c642b9" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.795019] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Updated the network info_cache for instance {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 740.795019] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Acquired lock "refresh_cache-915c3b6c-640b-430f-b264-40bf85c642b9" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.795019] env[61970]: DEBUG nova.network.neutron [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 740.795468] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 740.795639] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 740.795785] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 740.795931] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 740.796111] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 740.796264] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 740.796390] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61970) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 740.796526] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 741.140769] env[61970]: DEBUG oslo_concurrency.lockutils [req-b68911a8-73ea-442e-ba6b-405bdc220852 req-d4552b62-7b8e-40d3-be90-8405f3a29cab service nova] Releasing lock "refresh_cache-34719fba-dc16-4b43-ab48-517982ca85a4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.141036] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Acquired lock "refresh_cache-34719fba-dc16-4b43-ab48-517982ca85a4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.141233] env[61970]: DEBUG nova.network.neutron [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 741.212895] env[61970]: DEBUG nova.compute.utils [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 741.213970] env[61970]: DEBUG nova.compute.manager [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 741.213970] env[61970]: DEBUG nova.network.neutron [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 741.286111] env[61970]: DEBUG nova.policy [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '003e650d0730442ca780813683190422', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e0cb0afcc68342a18415c589150fb91f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 741.304445] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.321548] env[61970]: DEBUG nova.network.neutron [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.406683] env[61970]: DEBUG nova.network.neutron [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.643881] env[61970]: DEBUG nova.network.neutron [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Successfully created port: 242ccfc0-6f0b-43ae-b607-a09016baa251 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 741.666142] env[61970]: DEBUG nova.network.neutron [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.717372] env[61970]: DEBUG nova.compute.manager [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 741.810260] env[61970]: DEBUG nova.network.neutron [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.912661] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Releasing lock "refresh_cache-915c3b6c-640b-430f-b264-40bf85c642b9" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.915018] env[61970]: DEBUG nova.compute.manager [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 741.915018] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 741.915018] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1216b1a2-12ac-4e45-ae0a-22c2b114af5f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.921211] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 741.922316] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ae995d02-c85e-4d22-837f-0ede0f0956c8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.930804] env[61970]: DEBUG oslo_vmware.api [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 741.930804] env[61970]: value = "task-1355548" [ 741.930804] env[61970]: _type = "Task" [ 741.930804] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.940632] env[61970]: DEBUG oslo_vmware.api [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355548, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.992576] env[61970]: DEBUG nova.network.neutron [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Successfully created port: 6bdf88cb-0259-4814-8890-cf777f32cee5 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 742.079159] env[61970]: DEBUG nova.compute.manager [req-682d0331-c1cb-4528-bf91-11b439524a2d req-d0e46474-30cc-4614-a442-afa292271aa0 service nova] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Received event network-vif-deleted-663f1e4e-7249-473d-8538-7f4e66ba7d9c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 742.180960] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e9213cc-a9e0-4c79-a685-9ceeb788d8bd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.188667] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35b526d-37a3-4ff0-b064-1603d9a80481 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.218345] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478de8e8-ba69-47c8-b0d2-c981abb88722 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.229077] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76148967-c40c-4c63-a038-bb69647f67f2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.242590] env[61970]: DEBUG nova.compute.provider_tree [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.312569] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Releasing lock "refresh_cache-34719fba-dc16-4b43-ab48-517982ca85a4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.312972] env[61970]: DEBUG nova.compute.manager [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 742.313171] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 742.313483] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8e8e4ed6-a7ae-439e-9ad0-e457a51effd1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.323345] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9199b5bc-856b-47b7-998e-15a1ba44b490 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.346858] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 34719fba-dc16-4b43-ab48-517982ca85a4 could not be found. [ 742.347190] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 742.347270] env[61970]: INFO nova.compute.manager [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 742.347451] env[61970]: DEBUG oslo.service.loopingcall [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 742.347794] env[61970]: DEBUG nova.compute.manager [-] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 742.347794] env[61970]: DEBUG nova.network.neutron [-] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 742.365438] env[61970]: DEBUG nova.network.neutron [-] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.441046] env[61970]: DEBUG oslo_vmware.api [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355548, 'name': PowerOffVM_Task, 'duration_secs': 0.127806} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.441256] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 742.441417] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 742.441659] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e9f51523-d588-4056-bc4d-950ed58452b0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.466136] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 742.466443] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 742.467252] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Deleting the datastore file [datastore2] 915c3b6c-640b-430f-b264-40bf85c642b9 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 742.467252] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-07460f09-0e9f-4bfd-b58e-f7f0148d67d2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.472554] env[61970]: DEBUG oslo_vmware.api [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for the task: (returnval){ [ 742.472554] env[61970]: value = "task-1355550" [ 742.472554] env[61970]: _type = "Task" [ 742.472554] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.480423] env[61970]: DEBUG oslo_vmware.api [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355550, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.736623] env[61970]: DEBUG nova.compute.manager [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 742.745447] env[61970]: DEBUG nova.scheduler.client.report [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 742.770445] env[61970]: DEBUG nova.virt.hardware [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 742.770708] env[61970]: DEBUG nova.virt.hardware [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 742.770878] env[61970]: DEBUG nova.virt.hardware [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 742.771096] env[61970]: DEBUG nova.virt.hardware [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 742.772979] env[61970]: DEBUG nova.virt.hardware [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 742.772979] env[61970]: DEBUG nova.virt.hardware [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 742.772979] env[61970]: DEBUG nova.virt.hardware [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 742.772979] env[61970]: DEBUG nova.virt.hardware [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 742.772979] env[61970]: DEBUG nova.virt.hardware [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 742.773229] env[61970]: DEBUG nova.virt.hardware [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 742.773229] env[61970]: DEBUG nova.virt.hardware [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 742.773229] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9287ffd4-0c39-45fb-a8b9-6503a7157c94 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.781679] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2689f253-23bb-4d81-8028-84e0510b5584 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.869098] env[61970]: DEBUG nova.network.neutron [-] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.980706] env[61970]: ERROR nova.compute.manager [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 242ccfc0-6f0b-43ae-b607-a09016baa251, please check neutron logs for more information. [ 742.980706] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 742.980706] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 742.980706] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 742.980706] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.980706] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 742.980706] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.980706] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 742.980706] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.980706] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 742.980706] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.980706] env[61970]: ERROR nova.compute.manager raise self.value [ 742.980706] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.980706] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 742.980706] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.980706] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 742.981374] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.981374] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 742.981374] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 242ccfc0-6f0b-43ae-b607-a09016baa251, please check neutron logs for more information. [ 742.981374] env[61970]: ERROR nova.compute.manager [ 742.981374] env[61970]: Traceback (most recent call last): [ 742.981374] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 742.981374] env[61970]: listener.cb(fileno) [ 742.981374] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.981374] env[61970]: result = function(*args, **kwargs) [ 742.981374] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 742.981374] env[61970]: return func(*args, **kwargs) [ 742.981374] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 742.981374] env[61970]: raise e [ 742.981374] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 742.981374] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 742.981374] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.981374] env[61970]: created_port_ids = self._update_ports_for_instance( [ 742.981374] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.981374] env[61970]: with excutils.save_and_reraise_exception(): [ 742.981374] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.981374] env[61970]: self.force_reraise() [ 742.981374] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.981374] env[61970]: raise self.value [ 742.981374] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.981374] env[61970]: updated_port = self._update_port( [ 742.981374] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.981374] env[61970]: _ensure_no_port_binding_failure(port) [ 742.981374] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.981374] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 742.982324] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 242ccfc0-6f0b-43ae-b607-a09016baa251, please check neutron logs for more information. [ 742.982324] env[61970]: Removing descriptor: 15 [ 742.982796] env[61970]: ERROR nova.compute.manager [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 242ccfc0-6f0b-43ae-b607-a09016baa251, please check neutron logs for more information. [ 742.982796] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Traceback (most recent call last): [ 742.982796] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 742.982796] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] yield resources [ 742.982796] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 742.982796] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] self.driver.spawn(context, instance, image_meta, [ 742.982796] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 742.982796] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] self._vmops.spawn(context, instance, image_meta, injected_files, [ 742.982796] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 742.982796] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] vm_ref = self.build_virtual_machine(instance, [ 742.982796] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 742.983174] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] vif_infos = vmwarevif.get_vif_info(self._session, [ 742.983174] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 742.983174] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] for vif in network_info: [ 742.983174] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 742.983174] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] return self._sync_wrapper(fn, *args, **kwargs) [ 742.983174] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 742.983174] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] self.wait() [ 742.983174] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 742.983174] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] self[:] = self._gt.wait() [ 742.983174] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 742.983174] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] return self._exit_event.wait() [ 742.983174] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 742.983174] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] result = hub.switch() [ 742.983608] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 742.983608] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] return self.greenlet.switch() [ 742.983608] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.983608] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] result = function(*args, **kwargs) [ 742.983608] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 742.983608] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] return func(*args, **kwargs) [ 742.983608] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 742.983608] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] raise e [ 742.983608] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 742.983608] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] nwinfo = self.network_api.allocate_for_instance( [ 742.983608] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.983608] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] created_port_ids = self._update_ports_for_instance( [ 742.983608] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.984056] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] with excutils.save_and_reraise_exception(): [ 742.984056] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.984056] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] self.force_reraise() [ 742.984056] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.984056] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] raise self.value [ 742.984056] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.984056] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] updated_port = self._update_port( [ 742.984056] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.984056] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] _ensure_no_port_binding_failure(port) [ 742.984056] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.984056] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] raise exception.PortBindingFailed(port_id=port['id']) [ 742.984056] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] nova.exception.PortBindingFailed: Binding failed for port 242ccfc0-6f0b-43ae-b607-a09016baa251, please check neutron logs for more information. [ 742.984056] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] [ 742.984486] env[61970]: INFO nova.compute.manager [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Terminating instance [ 742.987342] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Acquiring lock "refresh_cache-1dc83f19-2965-418b-8d2a-56aa1124d258" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.987665] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Acquired lock "refresh_cache-1dc83f19-2965-418b-8d2a-56aa1124d258" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.987921] env[61970]: DEBUG nova.network.neutron [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 742.992509] env[61970]: DEBUG oslo_vmware.api [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Task: {'id': task-1355550, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.110715} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.993028] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 742.993314] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 742.993586] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 742.993843] env[61970]: INFO nova.compute.manager [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Took 1.08 seconds to destroy the instance on the hypervisor. [ 742.994216] env[61970]: DEBUG oslo.service.loopingcall [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 742.994495] env[61970]: DEBUG nova.compute.manager [-] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 742.994658] env[61970]: DEBUG nova.network.neutron [-] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 743.010481] env[61970]: DEBUG nova.network.neutron [-] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.250327] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.540s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.250921] env[61970]: DEBUG nova.compute.manager [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 743.253453] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.953s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.372173] env[61970]: INFO nova.compute.manager [-] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Took 1.02 seconds to deallocate network for instance. [ 743.374841] env[61970]: DEBUG nova.compute.claims [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 743.375036] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.512678] env[61970]: DEBUG nova.network.neutron [-] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.528343] env[61970]: DEBUG nova.network.neutron [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.613342] env[61970]: DEBUG nova.network.neutron [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.758204] env[61970]: DEBUG nova.compute.utils [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 743.763490] env[61970]: DEBUG nova.compute.manager [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 743.763677] env[61970]: DEBUG nova.network.neutron [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 743.962062] env[61970]: DEBUG nova.policy [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b277264800694382b457adf92b3e0c2e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8614d09e2b7e4666b717453d23394997', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 744.015478] env[61970]: INFO nova.compute.manager [-] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Took 1.02 seconds to deallocate network for instance. [ 744.108837] env[61970]: DEBUG nova.compute.manager [req-dde3cf34-9898-40b8-977a-88c8adacedea req-8dcb468f-2fed-48c3-bf1a-ec3d6af298aa service nova] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Received event network-changed-242ccfc0-6f0b-43ae-b607-a09016baa251 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 744.109085] env[61970]: DEBUG nova.compute.manager [req-dde3cf34-9898-40b8-977a-88c8adacedea req-8dcb468f-2fed-48c3-bf1a-ec3d6af298aa service nova] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Refreshing instance network info cache due to event network-changed-242ccfc0-6f0b-43ae-b607-a09016baa251. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 744.109239] env[61970]: DEBUG oslo_concurrency.lockutils [req-dde3cf34-9898-40b8-977a-88c8adacedea req-8dcb468f-2fed-48c3-bf1a-ec3d6af298aa service nova] Acquiring lock "refresh_cache-1dc83f19-2965-418b-8d2a-56aa1124d258" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.116927] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Releasing lock "refresh_cache-1dc83f19-2965-418b-8d2a-56aa1124d258" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.117448] env[61970]: DEBUG nova.compute.manager [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 744.117724] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 744.118129] env[61970]: DEBUG oslo_concurrency.lockutils [req-dde3cf34-9898-40b8-977a-88c8adacedea req-8dcb468f-2fed-48c3-bf1a-ec3d6af298aa service nova] Acquired lock "refresh_cache-1dc83f19-2965-418b-8d2a-56aa1124d258" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.118371] env[61970]: DEBUG nova.network.neutron [req-dde3cf34-9898-40b8-977a-88c8adacedea req-8dcb468f-2fed-48c3-bf1a-ec3d6af298aa service nova] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Refreshing network info cache for port 242ccfc0-6f0b-43ae-b607-a09016baa251 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 744.120587] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8484cb21-07fd-49ce-8c50-25a279d6ab67 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.139225] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a6e329-b7f2-44a6-ade0-8014f7558d9d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.152823] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c0033d-cff8-4f8e-b9eb-a14e5da6ea16 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.162337] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02d9545b-feaf-4e47-a718-63aec7daf09f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.177656] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1dc83f19-2965-418b-8d2a-56aa1124d258 could not be found. [ 744.180625] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 744.180951] env[61970]: INFO nova.compute.manager [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Took 0.06 seconds to destroy the instance on the hypervisor. [ 744.181289] env[61970]: DEBUG oslo.service.loopingcall [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 744.182007] env[61970]: DEBUG nova.compute.manager [-] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 744.182128] env[61970]: DEBUG nova.network.neutron [-] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 744.212360] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a8247b-0253-4161-bb20-72fbcc878d96 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.219646] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf33ea18-5efb-41d8-8d7e-fa79c6b2ca76 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.232730] env[61970]: DEBUG nova.compute.provider_tree [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.243917] env[61970]: DEBUG nova.network.neutron [-] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.264493] env[61970]: DEBUG nova.compute.manager [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 744.372964] env[61970]: DEBUG nova.network.neutron [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Successfully created port: 4026ce5e-c59e-459c-b80b-1ee765d383bf {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 744.531662] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.660697] env[61970]: DEBUG nova.network.neutron [req-dde3cf34-9898-40b8-977a-88c8adacedea req-8dcb468f-2fed-48c3-bf1a-ec3d6af298aa service nova] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.736595] env[61970]: DEBUG nova.scheduler.client.report [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 744.828393] env[61970]: DEBUG nova.network.neutron [req-dde3cf34-9898-40b8-977a-88c8adacedea req-8dcb468f-2fed-48c3-bf1a-ec3d6af298aa service nova] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.106095] env[61970]: DEBUG nova.network.neutron [-] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.247690] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.994s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.248412] env[61970]: ERROR nova.compute.manager [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 65286ee8-9e57-467d-887e-f5cfacaa25e5, please check neutron logs for more information. [ 745.248412] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Traceback (most recent call last): [ 745.248412] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 745.248412] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] self.driver.spawn(context, instance, image_meta, [ 745.248412] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 745.248412] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 745.248412] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 745.248412] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] vm_ref = self.build_virtual_machine(instance, [ 745.248412] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 745.248412] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 745.248412] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 745.248780] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] for vif in network_info: [ 745.248780] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 745.248780] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] return self._sync_wrapper(fn, *args, **kwargs) [ 745.248780] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 745.248780] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] self.wait() [ 745.248780] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 745.248780] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] self[:] = self._gt.wait() [ 745.248780] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 745.248780] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] return self._exit_event.wait() [ 745.248780] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 745.248780] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] result = hub.switch() [ 745.248780] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 745.248780] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] return self.greenlet.switch() [ 745.249207] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 745.249207] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] result = function(*args, **kwargs) [ 745.249207] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 745.249207] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] return func(*args, **kwargs) [ 745.249207] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 745.249207] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] raise e [ 745.249207] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.249207] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] nwinfo = self.network_api.allocate_for_instance( [ 745.249207] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 745.249207] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] created_port_ids = self._update_ports_for_instance( [ 745.249207] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 745.249207] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] with excutils.save_and_reraise_exception(): [ 745.249207] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.249598] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] self.force_reraise() [ 745.249598] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.249598] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] raise self.value [ 745.249598] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 745.249598] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] updated_port = self._update_port( [ 745.249598] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.249598] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] _ensure_no_port_binding_failure(port) [ 745.249598] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.249598] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] raise exception.PortBindingFailed(port_id=port['id']) [ 745.249598] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] nova.exception.PortBindingFailed: Binding failed for port 65286ee8-9e57-467d-887e-f5cfacaa25e5, please check neutron logs for more information. [ 745.249598] env[61970]: ERROR nova.compute.manager [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] [ 745.249954] env[61970]: DEBUG nova.compute.utils [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Binding failed for port 65286ee8-9e57-467d-887e-f5cfacaa25e5, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 745.250984] env[61970]: DEBUG oslo_concurrency.lockutils [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.378s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.252481] env[61970]: INFO nova.compute.claims [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 745.255381] env[61970]: DEBUG nova.compute.manager [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Build of instance 1a9257f6-5493-4ebb-8c37-88e11480a4f6 was re-scheduled: Binding failed for port 65286ee8-9e57-467d-887e-f5cfacaa25e5, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 745.256028] env[61970]: DEBUG nova.compute.manager [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 745.256028] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Acquiring lock "refresh_cache-1a9257f6-5493-4ebb-8c37-88e11480a4f6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.256151] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Acquired lock "refresh_cache-1a9257f6-5493-4ebb-8c37-88e11480a4f6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.256370] env[61970]: DEBUG nova.network.neutron [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 745.278288] env[61970]: DEBUG nova.compute.manager [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 745.317359] env[61970]: DEBUG nova.virt.hardware [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 745.317613] env[61970]: DEBUG nova.virt.hardware [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 745.317768] env[61970]: DEBUG nova.virt.hardware [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 745.317945] env[61970]: DEBUG nova.virt.hardware [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 745.318103] env[61970]: DEBUG nova.virt.hardware [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 745.318251] env[61970]: DEBUG nova.virt.hardware [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 745.318458] env[61970]: DEBUG nova.virt.hardware [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 745.318617] env[61970]: DEBUG nova.virt.hardware [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 745.318813] env[61970]: DEBUG nova.virt.hardware [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 745.318930] env[61970]: DEBUG nova.virt.hardware [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 745.319114] env[61970]: DEBUG nova.virt.hardware [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 745.320297] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a736a009-e57f-4abe-a8da-4673f951adaf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.328664] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1efba81d-340f-4cc4-8320-ff3027b384ca {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.332765] env[61970]: DEBUG oslo_concurrency.lockutils [req-dde3cf34-9898-40b8-977a-88c8adacedea req-8dcb468f-2fed-48c3-bf1a-ec3d6af298aa service nova] Releasing lock "refresh_cache-1dc83f19-2965-418b-8d2a-56aa1124d258" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.333081] env[61970]: DEBUG nova.compute.manager [req-dde3cf34-9898-40b8-977a-88c8adacedea req-8dcb468f-2fed-48c3-bf1a-ec3d6af298aa service nova] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Received event network-vif-deleted-242ccfc0-6f0b-43ae-b607-a09016baa251 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 745.562252] env[61970]: ERROR nova.compute.manager [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4026ce5e-c59e-459c-b80b-1ee765d383bf, please check neutron logs for more information. [ 745.562252] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 745.562252] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.562252] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 745.562252] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 745.562252] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 745.562252] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 745.562252] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 745.562252] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.562252] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 745.562252] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.562252] env[61970]: ERROR nova.compute.manager raise self.value [ 745.562252] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 745.562252] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 745.562252] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.562252] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 745.562712] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.562712] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 745.562712] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4026ce5e-c59e-459c-b80b-1ee765d383bf, please check neutron logs for more information. [ 745.562712] env[61970]: ERROR nova.compute.manager [ 745.562712] env[61970]: Traceback (most recent call last): [ 745.562712] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 745.562712] env[61970]: listener.cb(fileno) [ 745.562712] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 745.562712] env[61970]: result = function(*args, **kwargs) [ 745.562712] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 745.562712] env[61970]: return func(*args, **kwargs) [ 745.562712] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 745.562712] env[61970]: raise e [ 745.562712] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.562712] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 745.562712] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 745.562712] env[61970]: created_port_ids = self._update_ports_for_instance( [ 745.562712] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 745.562712] env[61970]: with excutils.save_and_reraise_exception(): [ 745.562712] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.562712] env[61970]: self.force_reraise() [ 745.562712] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.562712] env[61970]: raise self.value [ 745.562712] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 745.562712] env[61970]: updated_port = self._update_port( [ 745.562712] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.562712] env[61970]: _ensure_no_port_binding_failure(port) [ 745.562712] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.562712] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 745.563585] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 4026ce5e-c59e-459c-b80b-1ee765d383bf, please check neutron logs for more information. [ 745.563585] env[61970]: Removing descriptor: 15 [ 745.563585] env[61970]: ERROR nova.compute.manager [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4026ce5e-c59e-459c-b80b-1ee765d383bf, please check neutron logs for more information. [ 745.563585] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Traceback (most recent call last): [ 745.563585] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 745.563585] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] yield resources [ 745.563585] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 745.563585] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] self.driver.spawn(context, instance, image_meta, [ 745.563585] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 745.563585] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] self._vmops.spawn(context, instance, image_meta, injected_files, [ 745.563585] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 745.563585] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] vm_ref = self.build_virtual_machine(instance, [ 745.563985] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 745.563985] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] vif_infos = vmwarevif.get_vif_info(self._session, [ 745.563985] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 745.563985] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] for vif in network_info: [ 745.563985] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 745.563985] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] return self._sync_wrapper(fn, *args, **kwargs) [ 745.563985] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 745.563985] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] self.wait() [ 745.563985] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 745.563985] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] self[:] = self._gt.wait() [ 745.563985] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 745.563985] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] return self._exit_event.wait() [ 745.563985] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 745.564412] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] result = hub.switch() [ 745.564412] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 745.564412] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] return self.greenlet.switch() [ 745.564412] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 745.564412] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] result = function(*args, **kwargs) [ 745.564412] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 745.564412] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] return func(*args, **kwargs) [ 745.564412] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 745.564412] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] raise e [ 745.564412] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.564412] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] nwinfo = self.network_api.allocate_for_instance( [ 745.564412] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 745.564412] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] created_port_ids = self._update_ports_for_instance( [ 745.564833] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 745.564833] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] with excutils.save_and_reraise_exception(): [ 745.564833] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.564833] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] self.force_reraise() [ 745.564833] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.564833] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] raise self.value [ 745.564833] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 745.564833] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] updated_port = self._update_port( [ 745.564833] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.564833] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] _ensure_no_port_binding_failure(port) [ 745.564833] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.564833] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] raise exception.PortBindingFailed(port_id=port['id']) [ 745.565255] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] nova.exception.PortBindingFailed: Binding failed for port 4026ce5e-c59e-459c-b80b-1ee765d383bf, please check neutron logs for more information. [ 745.565255] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] [ 745.565255] env[61970]: INFO nova.compute.manager [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Terminating instance [ 745.565603] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Acquiring lock "refresh_cache-ed10e0a6-c79f-4783-bb93-312c5f0ffd51" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.565763] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Acquired lock "refresh_cache-ed10e0a6-c79f-4783-bb93-312c5f0ffd51" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.565929] env[61970]: DEBUG nova.network.neutron [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 745.609050] env[61970]: INFO nova.compute.manager [-] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Took 1.43 seconds to deallocate network for instance. [ 745.611156] env[61970]: DEBUG nova.compute.claims [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 745.611328] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.771522] env[61970]: DEBUG nova.network.neutron [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.815585] env[61970]: DEBUG nova.network.neutron [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.089998] env[61970]: DEBUG nova.network.neutron [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.178309] env[61970]: DEBUG nova.compute.manager [req-8aefcde1-cc97-4f00-8f24-68de0aa06635 req-8514448a-4d7e-4436-b471-7b8eb3b774f1 service nova] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Received event network-changed-4026ce5e-c59e-459c-b80b-1ee765d383bf {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 746.178508] env[61970]: DEBUG nova.compute.manager [req-8aefcde1-cc97-4f00-8f24-68de0aa06635 req-8514448a-4d7e-4436-b471-7b8eb3b774f1 service nova] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Refreshing instance network info cache due to event network-changed-4026ce5e-c59e-459c-b80b-1ee765d383bf. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 746.178710] env[61970]: DEBUG oslo_concurrency.lockutils [req-8aefcde1-cc97-4f00-8f24-68de0aa06635 req-8514448a-4d7e-4436-b471-7b8eb3b774f1 service nova] Acquiring lock "refresh_cache-ed10e0a6-c79f-4783-bb93-312c5f0ffd51" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.179712] env[61970]: DEBUG nova.network.neutron [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.318134] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Releasing lock "refresh_cache-1a9257f6-5493-4ebb-8c37-88e11480a4f6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.318365] env[61970]: DEBUG nova.compute.manager [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 746.318530] env[61970]: DEBUG nova.compute.manager [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 746.318693] env[61970]: DEBUG nova.network.neutron [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 746.333853] env[61970]: DEBUG nova.network.neutron [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.557752] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0db4c533-9788-46da-bbff-df0525c71847 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.565778] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfa01c4-0230-4ebc-a9b5-12512c0561e1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.595584] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1424c6d1-0603-49c5-a69a-e2aeb98b161a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.602899] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c26cab-f414-4d31-ac42-9d94a72326f9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.616187] env[61970]: DEBUG nova.compute.provider_tree [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.682574] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Releasing lock "refresh_cache-ed10e0a6-c79f-4783-bb93-312c5f0ffd51" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.683095] env[61970]: DEBUG nova.compute.manager [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 746.683095] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 746.683348] env[61970]: DEBUG oslo_concurrency.lockutils [req-8aefcde1-cc97-4f00-8f24-68de0aa06635 req-8514448a-4d7e-4436-b471-7b8eb3b774f1 service nova] Acquired lock "refresh_cache-ed10e0a6-c79f-4783-bb93-312c5f0ffd51" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.684326] env[61970]: DEBUG nova.network.neutron [req-8aefcde1-cc97-4f00-8f24-68de0aa06635 req-8514448a-4d7e-4436-b471-7b8eb3b774f1 service nova] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Refreshing network info cache for port 4026ce5e-c59e-459c-b80b-1ee765d383bf {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 746.684592] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7fbfdcad-17f7-49d4-8cd1-5739392092b9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.696049] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae96c385-9226-47c2-887a-1b12ca64aa7d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.718333] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ed10e0a6-c79f-4783-bb93-312c5f0ffd51 could not be found. [ 746.718565] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 746.718745] env[61970]: INFO nova.compute.manager [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Took 0.04 seconds to destroy the instance on the hypervisor. [ 746.718991] env[61970]: DEBUG oslo.service.loopingcall [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 746.719478] env[61970]: DEBUG nova.compute.manager [-] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 746.719587] env[61970]: DEBUG nova.network.neutron [-] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 746.733453] env[61970]: DEBUG nova.network.neutron [-] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.836619] env[61970]: DEBUG nova.network.neutron [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.119440] env[61970]: DEBUG nova.scheduler.client.report [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 747.212263] env[61970]: DEBUG nova.network.neutron [req-8aefcde1-cc97-4f00-8f24-68de0aa06635 req-8514448a-4d7e-4436-b471-7b8eb3b774f1 service nova] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.235978] env[61970]: DEBUG nova.network.neutron [-] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.292176] env[61970]: DEBUG nova.network.neutron [req-8aefcde1-cc97-4f00-8f24-68de0aa06635 req-8514448a-4d7e-4436-b471-7b8eb3b774f1 service nova] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.340896] env[61970]: INFO nova.compute.manager [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] [instance: 1a9257f6-5493-4ebb-8c37-88e11480a4f6] Took 1.02 seconds to deallocate network for instance. [ 747.624245] env[61970]: DEBUG oslo_concurrency.lockutils [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.624775] env[61970]: DEBUG nova.compute.manager [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 747.627573] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.782s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.738815] env[61970]: INFO nova.compute.manager [-] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Took 1.02 seconds to deallocate network for instance. [ 747.741148] env[61970]: DEBUG nova.compute.claims [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 747.741342] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.795031] env[61970]: DEBUG oslo_concurrency.lockutils [req-8aefcde1-cc97-4f00-8f24-68de0aa06635 req-8514448a-4d7e-4436-b471-7b8eb3b774f1 service nova] Releasing lock "refresh_cache-ed10e0a6-c79f-4783-bb93-312c5f0ffd51" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.795031] env[61970]: DEBUG nova.compute.manager [req-8aefcde1-cc97-4f00-8f24-68de0aa06635 req-8514448a-4d7e-4436-b471-7b8eb3b774f1 service nova] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Received event network-vif-deleted-4026ce5e-c59e-459c-b80b-1ee765d383bf {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 748.135299] env[61970]: DEBUG nova.compute.utils [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 748.137624] env[61970]: DEBUG nova.compute.manager [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 748.137869] env[61970]: DEBUG nova.network.neutron [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 748.176898] env[61970]: DEBUG nova.policy [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b277264800694382b457adf92b3e0c2e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8614d09e2b7e4666b717453d23394997', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 748.376528] env[61970]: INFO nova.scheduler.client.report [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Deleted allocations for instance 1a9257f6-5493-4ebb-8c37-88e11480a4f6 [ 748.460811] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d101398b-cd65-43b0-926c-2480f73e8ed9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.469026] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56410d25-c12e-480a-8867-8ec623692916 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.500302] env[61970]: DEBUG nova.network.neutron [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Successfully created port: 35fe4568-400f-49b0-8ad9-0925de026db3 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 748.502722] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d3f0b6-13bb-41ee-adde-d69f238e6898 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.510191] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7579a091-a027-4d06-a36b-86f65d6b0af4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.528019] env[61970]: DEBUG nova.compute.provider_tree [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.641033] env[61970]: DEBUG nova.compute.manager [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 748.886302] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1732a64d-b929-4af1-a59c-f52d75a0bc20 tempest-ServerDiagnosticsNegativeTest-1336132834 tempest-ServerDiagnosticsNegativeTest-1336132834-project-member] Lock "1a9257f6-5493-4ebb-8c37-88e11480a4f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.723s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.029039] env[61970]: DEBUG nova.scheduler.client.report [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 749.304896] env[61970]: DEBUG nova.compute.manager [req-84f86d3b-ce19-4345-9839-ea90ccab630b req-a3fc2949-40c0-461a-98ca-8482b5edd945 service nova] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Received event network-changed-35fe4568-400f-49b0-8ad9-0925de026db3 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 749.305177] env[61970]: DEBUG nova.compute.manager [req-84f86d3b-ce19-4345-9839-ea90ccab630b req-a3fc2949-40c0-461a-98ca-8482b5edd945 service nova] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Refreshing instance network info cache due to event network-changed-35fe4568-400f-49b0-8ad9-0925de026db3. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 749.305711] env[61970]: DEBUG oslo_concurrency.lockutils [req-84f86d3b-ce19-4345-9839-ea90ccab630b req-a3fc2949-40c0-461a-98ca-8482b5edd945 service nova] Acquiring lock "refresh_cache-5d352f39-4544-46ec-af43-e539d511b47b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.305856] env[61970]: DEBUG oslo_concurrency.lockutils [req-84f86d3b-ce19-4345-9839-ea90ccab630b req-a3fc2949-40c0-461a-98ca-8482b5edd945 service nova] Acquired lock "refresh_cache-5d352f39-4544-46ec-af43-e539d511b47b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.306116] env[61970]: DEBUG nova.network.neutron [req-84f86d3b-ce19-4345-9839-ea90ccab630b req-a3fc2949-40c0-461a-98ca-8482b5edd945 service nova] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Refreshing network info cache for port 35fe4568-400f-49b0-8ad9-0925de026db3 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 749.389686] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 749.514798] env[61970]: ERROR nova.compute.manager [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 35fe4568-400f-49b0-8ad9-0925de026db3, please check neutron logs for more information. [ 749.514798] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 749.514798] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.514798] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 749.514798] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.514798] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 749.514798] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.514798] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 749.514798] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.514798] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 749.514798] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.514798] env[61970]: ERROR nova.compute.manager raise self.value [ 749.514798] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.514798] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 749.514798] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.514798] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 749.515331] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.515331] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 749.515331] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 35fe4568-400f-49b0-8ad9-0925de026db3, please check neutron logs for more information. [ 749.515331] env[61970]: ERROR nova.compute.manager [ 749.515331] env[61970]: Traceback (most recent call last): [ 749.515331] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 749.515331] env[61970]: listener.cb(fileno) [ 749.515331] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.515331] env[61970]: result = function(*args, **kwargs) [ 749.515331] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 749.515331] env[61970]: return func(*args, **kwargs) [ 749.515331] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.515331] env[61970]: raise e [ 749.515331] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.515331] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 749.515331] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.515331] env[61970]: created_port_ids = self._update_ports_for_instance( [ 749.515331] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.515331] env[61970]: with excutils.save_and_reraise_exception(): [ 749.515331] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.515331] env[61970]: self.force_reraise() [ 749.515331] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.515331] env[61970]: raise self.value [ 749.515331] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.515331] env[61970]: updated_port = self._update_port( [ 749.515331] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.515331] env[61970]: _ensure_no_port_binding_failure(port) [ 749.515331] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.515331] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 749.516131] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 35fe4568-400f-49b0-8ad9-0925de026db3, please check neutron logs for more information. [ 749.516131] env[61970]: Removing descriptor: 15 [ 749.535052] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.908s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.535659] env[61970]: ERROR nova.compute.manager [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 091e50a0-7fd8-4d7c-9035-4a29f32819ae, please check neutron logs for more information. [ 749.535659] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Traceback (most recent call last): [ 749.535659] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 749.535659] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] self.driver.spawn(context, instance, image_meta, [ 749.535659] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 749.535659] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.535659] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.535659] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] vm_ref = self.build_virtual_machine(instance, [ 749.535659] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.535659] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.535659] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.536196] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] for vif in network_info: [ 749.536196] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 749.536196] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] return self._sync_wrapper(fn, *args, **kwargs) [ 749.536196] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 749.536196] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] self.wait() [ 749.536196] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 749.536196] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] self[:] = self._gt.wait() [ 749.536196] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.536196] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] return self._exit_event.wait() [ 749.536196] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 749.536196] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] current.throw(*self._exc) [ 749.536196] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.536196] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] result = function(*args, **kwargs) [ 749.536856] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 749.536856] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] return func(*args, **kwargs) [ 749.536856] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.536856] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] raise e [ 749.536856] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.536856] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] nwinfo = self.network_api.allocate_for_instance( [ 749.536856] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.536856] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] created_port_ids = self._update_ports_for_instance( [ 749.536856] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.536856] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] with excutils.save_and_reraise_exception(): [ 749.536856] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.536856] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] self.force_reraise() [ 749.536856] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.537540] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] raise self.value [ 749.537540] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.537540] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] updated_port = self._update_port( [ 749.537540] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.537540] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] _ensure_no_port_binding_failure(port) [ 749.537540] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.537540] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] raise exception.PortBindingFailed(port_id=port['id']) [ 749.537540] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] nova.exception.PortBindingFailed: Binding failed for port 091e50a0-7fd8-4d7c-9035-4a29f32819ae, please check neutron logs for more information. [ 749.537540] env[61970]: ERROR nova.compute.manager [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] [ 749.537540] env[61970]: DEBUG nova.compute.utils [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Binding failed for port 091e50a0-7fd8-4d7c-9035-4a29f32819ae, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 749.538024] env[61970]: DEBUG oslo_concurrency.lockutils [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.086s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.539209] env[61970]: INFO nova.compute.claims [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 749.541980] env[61970]: DEBUG nova.compute.manager [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Build of instance 762efd5e-c95e-4718-98de-2e0b05226c06 was re-scheduled: Binding failed for port 091e50a0-7fd8-4d7c-9035-4a29f32819ae, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 749.542511] env[61970]: DEBUG nova.compute.manager [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 749.542748] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Acquiring lock "refresh_cache-762efd5e-c95e-4718-98de-2e0b05226c06" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.543175] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Acquired lock "refresh_cache-762efd5e-c95e-4718-98de-2e0b05226c06" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.543359] env[61970]: DEBUG nova.network.neutron [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 749.648670] env[61970]: DEBUG nova.compute.manager [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 749.672643] env[61970]: DEBUG nova.virt.hardware [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 749.672893] env[61970]: DEBUG nova.virt.hardware [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 749.673139] env[61970]: DEBUG nova.virt.hardware [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 749.673348] env[61970]: DEBUG nova.virt.hardware [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 749.673497] env[61970]: DEBUG nova.virt.hardware [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 749.673643] env[61970]: DEBUG nova.virt.hardware [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 749.673848] env[61970]: DEBUG nova.virt.hardware [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 749.674037] env[61970]: DEBUG nova.virt.hardware [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 749.674237] env[61970]: DEBUG nova.virt.hardware [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 749.674404] env[61970]: DEBUG nova.virt.hardware [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 749.674575] env[61970]: DEBUG nova.virt.hardware [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 749.675445] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373e5bc4-e6d4-4a93-b541-d55aa8f04c9d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.683841] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24030ab3-bb65-4799-9e6e-2bdb4b022daa {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.697938] env[61970]: ERROR nova.compute.manager [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 35fe4568-400f-49b0-8ad9-0925de026db3, please check neutron logs for more information. [ 749.697938] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Traceback (most recent call last): [ 749.697938] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 749.697938] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] yield resources [ 749.697938] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 749.697938] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] self.driver.spawn(context, instance, image_meta, [ 749.697938] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 749.697938] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.697938] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.697938] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] vm_ref = self.build_virtual_machine(instance, [ 749.697938] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.698365] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.698365] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.698365] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] for vif in network_info: [ 749.698365] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 749.698365] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] return self._sync_wrapper(fn, *args, **kwargs) [ 749.698365] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 749.698365] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] self.wait() [ 749.698365] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 749.698365] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] self[:] = self._gt.wait() [ 749.698365] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.698365] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] return self._exit_event.wait() [ 749.698365] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 749.698365] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] current.throw(*self._exc) [ 749.698816] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.698816] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] result = function(*args, **kwargs) [ 749.698816] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 749.698816] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] return func(*args, **kwargs) [ 749.698816] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.698816] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] raise e [ 749.698816] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.698816] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] nwinfo = self.network_api.allocate_for_instance( [ 749.698816] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.698816] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] created_port_ids = self._update_ports_for_instance( [ 749.698816] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.698816] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] with excutils.save_and_reraise_exception(): [ 749.698816] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.699246] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] self.force_reraise() [ 749.699246] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.699246] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] raise self.value [ 749.699246] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.699246] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] updated_port = self._update_port( [ 749.699246] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.699246] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] _ensure_no_port_binding_failure(port) [ 749.699246] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.699246] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] raise exception.PortBindingFailed(port_id=port['id']) [ 749.699246] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] nova.exception.PortBindingFailed: Binding failed for port 35fe4568-400f-49b0-8ad9-0925de026db3, please check neutron logs for more information. [ 749.699246] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] [ 749.699246] env[61970]: INFO nova.compute.manager [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Terminating instance [ 749.700171] env[61970]: DEBUG oslo_concurrency.lockutils [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Acquiring lock "refresh_cache-5d352f39-4544-46ec-af43-e539d511b47b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.824982] env[61970]: DEBUG nova.network.neutron [req-84f86d3b-ce19-4345-9839-ea90ccab630b req-a3fc2949-40c0-461a-98ca-8482b5edd945 service nova] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.912480] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.913347] env[61970]: DEBUG nova.network.neutron [req-84f86d3b-ce19-4345-9839-ea90ccab630b req-a3fc2949-40c0-461a-98ca-8482b5edd945 service nova] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.063468] env[61970]: DEBUG nova.network.neutron [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.137999] env[61970]: DEBUG nova.network.neutron [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.416208] env[61970]: DEBUG oslo_concurrency.lockutils [req-84f86d3b-ce19-4345-9839-ea90ccab630b req-a3fc2949-40c0-461a-98ca-8482b5edd945 service nova] Releasing lock "refresh_cache-5d352f39-4544-46ec-af43-e539d511b47b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.416474] env[61970]: DEBUG oslo_concurrency.lockutils [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Acquired lock "refresh_cache-5d352f39-4544-46ec-af43-e539d511b47b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.416651] env[61970]: DEBUG nova.network.neutron [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 750.641998] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Releasing lock "refresh_cache-762efd5e-c95e-4718-98de-2e0b05226c06" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.642260] env[61970]: DEBUG nova.compute.manager [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 750.642470] env[61970]: DEBUG nova.compute.manager [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 750.642635] env[61970]: DEBUG nova.network.neutron [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 750.676889] env[61970]: DEBUG nova.network.neutron [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.893826] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e429d8-9ed3-4766-b232-321bd995aabe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.900814] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61cd7d6e-cc38-47db-a51b-780c334c2cc3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.931477] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a56151-b1b6-4837-ae67-065268e866da {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.941090] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3296258f-f9ad-413c-a348-2a0b55c54b44 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.953081] env[61970]: DEBUG nova.compute.provider_tree [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.954952] env[61970]: DEBUG nova.network.neutron [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.053494] env[61970]: DEBUG nova.network.neutron [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.180422] env[61970]: DEBUG nova.network.neutron [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.332602] env[61970]: DEBUG nova.compute.manager [req-cf6190c2-a2f9-407d-88a9-bc5b62c1dac6 req-877ebefa-e3a1-4401-9bae-1ac5a6001601 service nova] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Received event network-vif-deleted-35fe4568-400f-49b0-8ad9-0925de026db3 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 751.459060] env[61970]: DEBUG nova.scheduler.client.report [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 751.556493] env[61970]: DEBUG oslo_concurrency.lockutils [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Releasing lock "refresh_cache-5d352f39-4544-46ec-af43-e539d511b47b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.556928] env[61970]: DEBUG nova.compute.manager [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 751.557275] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 751.557470] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c99c23df-d7bf-4e2f-865a-5fce363e0a65 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.566577] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d130d8-3e17-481d-9fa5-2318637edc62 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.587738] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5d352f39-4544-46ec-af43-e539d511b47b could not be found. [ 751.587956] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 751.588201] env[61970]: INFO nova.compute.manager [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 751.588464] env[61970]: DEBUG oslo.service.loopingcall [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.588685] env[61970]: DEBUG nova.compute.manager [-] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 751.588784] env[61970]: DEBUG nova.network.neutron [-] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 751.604630] env[61970]: DEBUG nova.network.neutron [-] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.683521] env[61970]: INFO nova.compute.manager [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] [instance: 762efd5e-c95e-4718-98de-2e0b05226c06] Took 1.04 seconds to deallocate network for instance. [ 751.963829] env[61970]: DEBUG oslo_concurrency.lockutils [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.426s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.964397] env[61970]: DEBUG nova.compute.manager [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 751.966899] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.384s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.967093] env[61970]: DEBUG nova.objects.instance [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61970) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 752.107066] env[61970]: DEBUG nova.network.neutron [-] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.471473] env[61970]: DEBUG nova.compute.utils [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 752.475386] env[61970]: DEBUG nova.compute.manager [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 752.476104] env[61970]: DEBUG nova.network.neutron [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 752.527934] env[61970]: DEBUG nova.policy [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '13aacb75013d408980ba5189c6071447', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c5b87b23cfe747379a4f42e6879f0f25', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 752.609534] env[61970]: INFO nova.compute.manager [-] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Took 1.02 seconds to deallocate network for instance. [ 752.611881] env[61970]: DEBUG nova.compute.claims [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 752.612102] env[61970]: DEBUG oslo_concurrency.lockutils [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.710654] env[61970]: INFO nova.scheduler.client.report [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Deleted allocations for instance 762efd5e-c95e-4718-98de-2e0b05226c06 [ 752.802642] env[61970]: DEBUG nova.network.neutron [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Successfully created port: 59301aa5-0ef4-43a3-87e6-1f968c41bd4f {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 752.975936] env[61970]: DEBUG nova.compute.manager [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 752.980207] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5cae8a4d-6904-49ee-96cf-39abc2022bb4 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.984019] env[61970]: DEBUG oslo_concurrency.lockutils [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.641s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.221166] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ca2c0a26-8c9a-4d6b-9f3c-132600a535d5 tempest-AttachInterfacesV270Test-798897830 tempest-AttachInterfacesV270Test-798897830-project-member] Lock "762efd5e-c95e-4718-98de-2e0b05226c06" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.111s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.723590] env[61970]: DEBUG nova.compute.manager [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 753.874830] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b4e2f8-7f76-4d08-acef-5b0f27e21da8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.883128] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4394d39a-f15e-48db-a2b3-3d7edee7649b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.921208] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b3ace43-13b3-4b95-8054-454ae5641e93 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.928756] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20b5f3f1-5b1b-4c25-9a05-9cab83580f4a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.944809] env[61970]: DEBUG nova.compute.provider_tree [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 753.992463] env[61970]: DEBUG nova.compute.manager [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 754.023839] env[61970]: DEBUG nova.virt.hardware [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 754.024110] env[61970]: DEBUG nova.virt.hardware [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 754.024303] env[61970]: DEBUG nova.virt.hardware [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 754.024514] env[61970]: DEBUG nova.virt.hardware [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 754.024662] env[61970]: DEBUG nova.virt.hardware [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 754.024807] env[61970]: DEBUG nova.virt.hardware [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 754.025031] env[61970]: DEBUG nova.virt.hardware [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 754.025196] env[61970]: DEBUG nova.virt.hardware [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 754.025361] env[61970]: DEBUG nova.virt.hardware [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 754.025521] env[61970]: DEBUG nova.virt.hardware [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 754.025712] env[61970]: DEBUG nova.virt.hardware [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 754.026922] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a49cb6-2da5-474a-8ffd-c3281d6cfbc4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.035897] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8332fe7-c10f-4f2d-8e36-c632cbfe94ad {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.177178] env[61970]: DEBUG nova.compute.manager [req-5ea1be59-7b7d-4206-b08c-56d9c2e3a898 req-bb1c7b34-0bd0-44ff-9428-53a850c85c08 service nova] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Received event network-changed-59301aa5-0ef4-43a3-87e6-1f968c41bd4f {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 754.177379] env[61970]: DEBUG nova.compute.manager [req-5ea1be59-7b7d-4206-b08c-56d9c2e3a898 req-bb1c7b34-0bd0-44ff-9428-53a850c85c08 service nova] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Refreshing instance network info cache due to event network-changed-59301aa5-0ef4-43a3-87e6-1f968c41bd4f. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 754.177591] env[61970]: DEBUG oslo_concurrency.lockutils [req-5ea1be59-7b7d-4206-b08c-56d9c2e3a898 req-bb1c7b34-0bd0-44ff-9428-53a850c85c08 service nova] Acquiring lock "refresh_cache-fd4274ed-bf9a-45ab-a350-a8ebbc5ad404" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.177736] env[61970]: DEBUG oslo_concurrency.lockutils [req-5ea1be59-7b7d-4206-b08c-56d9c2e3a898 req-bb1c7b34-0bd0-44ff-9428-53a850c85c08 service nova] Acquired lock "refresh_cache-fd4274ed-bf9a-45ab-a350-a8ebbc5ad404" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.177898] env[61970]: DEBUG nova.network.neutron [req-5ea1be59-7b7d-4206-b08c-56d9c2e3a898 req-bb1c7b34-0bd0-44ff-9428-53a850c85c08 service nova] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Refreshing network info cache for port 59301aa5-0ef4-43a3-87e6-1f968c41bd4f {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 754.249839] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.447967] env[61970]: DEBUG nova.scheduler.client.report [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.460578] env[61970]: ERROR nova.compute.manager [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 59301aa5-0ef4-43a3-87e6-1f968c41bd4f, please check neutron logs for more information. [ 754.460578] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 754.460578] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.460578] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 754.460578] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 754.460578] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 754.460578] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 754.460578] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 754.460578] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.460578] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 754.460578] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.460578] env[61970]: ERROR nova.compute.manager raise self.value [ 754.460578] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 754.460578] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 754.460578] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.460578] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 754.462506] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.462506] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 754.462506] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 59301aa5-0ef4-43a3-87e6-1f968c41bd4f, please check neutron logs for more information. [ 754.462506] env[61970]: ERROR nova.compute.manager [ 754.462506] env[61970]: Traceback (most recent call last): [ 754.462506] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 754.462506] env[61970]: listener.cb(fileno) [ 754.462506] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 754.462506] env[61970]: result = function(*args, **kwargs) [ 754.462506] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 754.462506] env[61970]: return func(*args, **kwargs) [ 754.462506] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 754.462506] env[61970]: raise e [ 754.462506] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.462506] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 754.462506] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 754.462506] env[61970]: created_port_ids = self._update_ports_for_instance( [ 754.462506] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 754.462506] env[61970]: with excutils.save_and_reraise_exception(): [ 754.462506] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.462506] env[61970]: self.force_reraise() [ 754.462506] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.462506] env[61970]: raise self.value [ 754.462506] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 754.462506] env[61970]: updated_port = self._update_port( [ 754.462506] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.462506] env[61970]: _ensure_no_port_binding_failure(port) [ 754.462506] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.462506] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 754.464617] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 59301aa5-0ef4-43a3-87e6-1f968c41bd4f, please check neutron logs for more information. [ 754.464617] env[61970]: Removing descriptor: 15 [ 754.464617] env[61970]: ERROR nova.compute.manager [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 59301aa5-0ef4-43a3-87e6-1f968c41bd4f, please check neutron logs for more information. [ 754.464617] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Traceback (most recent call last): [ 754.464617] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 754.464617] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] yield resources [ 754.464617] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 754.464617] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] self.driver.spawn(context, instance, image_meta, [ 754.464617] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 754.464617] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] self._vmops.spawn(context, instance, image_meta, injected_files, [ 754.464617] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 754.464617] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] vm_ref = self.build_virtual_machine(instance, [ 754.465101] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 754.465101] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] vif_infos = vmwarevif.get_vif_info(self._session, [ 754.465101] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 754.465101] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] for vif in network_info: [ 754.465101] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 754.465101] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] return self._sync_wrapper(fn, *args, **kwargs) [ 754.465101] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 754.465101] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] self.wait() [ 754.465101] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 754.465101] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] self[:] = self._gt.wait() [ 754.465101] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 754.465101] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] return self._exit_event.wait() [ 754.465101] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 754.465559] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] result = hub.switch() [ 754.465559] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 754.465559] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] return self.greenlet.switch() [ 754.465559] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 754.465559] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] result = function(*args, **kwargs) [ 754.465559] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 754.465559] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] return func(*args, **kwargs) [ 754.465559] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 754.465559] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] raise e [ 754.465559] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.465559] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] nwinfo = self.network_api.allocate_for_instance( [ 754.465559] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 754.465559] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] created_port_ids = self._update_ports_for_instance( [ 754.465983] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 754.465983] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] with excutils.save_and_reraise_exception(): [ 754.465983] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.465983] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] self.force_reraise() [ 754.465983] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.465983] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] raise self.value [ 754.465983] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 754.465983] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] updated_port = self._update_port( [ 754.465983] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.465983] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] _ensure_no_port_binding_failure(port) [ 754.465983] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.465983] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] raise exception.PortBindingFailed(port_id=port['id']) [ 754.466383] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] nova.exception.PortBindingFailed: Binding failed for port 59301aa5-0ef4-43a3-87e6-1f968c41bd4f, please check neutron logs for more information. [ 754.466383] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] [ 754.466383] env[61970]: INFO nova.compute.manager [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Terminating instance [ 754.466383] env[61970]: DEBUG oslo_concurrency.lockutils [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "refresh_cache-fd4274ed-bf9a-45ab-a350-a8ebbc5ad404" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.704610] env[61970]: DEBUG nova.network.neutron [req-5ea1be59-7b7d-4206-b08c-56d9c2e3a898 req-bb1c7b34-0bd0-44ff-9428-53a850c85c08 service nova] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.797589] env[61970]: DEBUG nova.network.neutron [req-5ea1be59-7b7d-4206-b08c-56d9c2e3a898 req-bb1c7b34-0bd0-44ff-9428-53a850c85c08 service nova] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.957087] env[61970]: DEBUG oslo_concurrency.lockutils [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.972s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.957087] env[61970]: ERROR nova.compute.manager [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6e1af134-e574-4f35-90e1-a639108500c5, please check neutron logs for more information. [ 754.957087] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Traceback (most recent call last): [ 754.957087] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 754.957087] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] self.driver.spawn(context, instance, image_meta, [ 754.957087] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 754.957087] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 754.957087] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 754.957087] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] vm_ref = self.build_virtual_machine(instance, [ 754.957569] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 754.957569] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 754.957569] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 754.957569] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] for vif in network_info: [ 754.957569] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 754.957569] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] return self._sync_wrapper(fn, *args, **kwargs) [ 754.957569] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 754.957569] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] self.wait() [ 754.957569] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 754.957569] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] self[:] = self._gt.wait() [ 754.957569] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 754.957569] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] return self._exit_event.wait() [ 754.957569] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 754.957993] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] current.throw(*self._exc) [ 754.957993] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 754.957993] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] result = function(*args, **kwargs) [ 754.957993] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 754.957993] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] return func(*args, **kwargs) [ 754.957993] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 754.957993] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] raise e [ 754.957993] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.957993] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] nwinfo = self.network_api.allocate_for_instance( [ 754.957993] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 754.957993] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] created_port_ids = self._update_ports_for_instance( [ 754.957993] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 754.957993] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] with excutils.save_and_reraise_exception(): [ 754.958429] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.958429] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] self.force_reraise() [ 754.958429] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.958429] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] raise self.value [ 754.958429] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 754.958429] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] updated_port = self._update_port( [ 754.958429] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.958429] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] _ensure_no_port_binding_failure(port) [ 754.958429] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.958429] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] raise exception.PortBindingFailed(port_id=port['id']) [ 754.958429] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] nova.exception.PortBindingFailed: Binding failed for port 6e1af134-e574-4f35-90e1-a639108500c5, please check neutron logs for more information. [ 754.958429] env[61970]: ERROR nova.compute.manager [instance: a540095e-563e-4059-b861-5d0e1b4995f4] [ 754.958822] env[61970]: DEBUG nova.compute.utils [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Binding failed for port 6e1af134-e574-4f35-90e1-a639108500c5, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 754.958822] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.029s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.958822] env[61970]: INFO nova.compute.claims [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 754.961236] env[61970]: DEBUG nova.compute.manager [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Build of instance a540095e-563e-4059-b861-5d0e1b4995f4 was re-scheduled: Binding failed for port 6e1af134-e574-4f35-90e1-a639108500c5, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 754.961728] env[61970]: DEBUG nova.compute.manager [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 754.961950] env[61970]: DEBUG oslo_concurrency.lockutils [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Acquiring lock "refresh_cache-a540095e-563e-4059-b861-5d0e1b4995f4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.962124] env[61970]: DEBUG oslo_concurrency.lockutils [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Acquired lock "refresh_cache-a540095e-563e-4059-b861-5d0e1b4995f4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.962281] env[61970]: DEBUG nova.network.neutron [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 755.301675] env[61970]: DEBUG oslo_concurrency.lockutils [req-5ea1be59-7b7d-4206-b08c-56d9c2e3a898 req-bb1c7b34-0bd0-44ff-9428-53a850c85c08 service nova] Releasing lock "refresh_cache-fd4274ed-bf9a-45ab-a350-a8ebbc5ad404" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.301781] env[61970]: DEBUG oslo_concurrency.lockutils [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquired lock "refresh_cache-fd4274ed-bf9a-45ab-a350-a8ebbc5ad404" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.303040] env[61970]: DEBUG nova.network.neutron [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 755.481761] env[61970]: DEBUG nova.network.neutron [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.559734] env[61970]: DEBUG nova.network.neutron [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.819900] env[61970]: DEBUG nova.network.neutron [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.905021] env[61970]: DEBUG nova.network.neutron [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.066020] env[61970]: DEBUG oslo_concurrency.lockutils [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Releasing lock "refresh_cache-a540095e-563e-4059-b861-5d0e1b4995f4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.066020] env[61970]: DEBUG nova.compute.manager [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 756.066020] env[61970]: DEBUG nova.compute.manager [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 756.066020] env[61970]: DEBUG nova.network.neutron [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 756.098735] env[61970]: DEBUG nova.network.neutron [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.213232] env[61970]: DEBUG nova.compute.manager [req-10162250-2346-417f-ad81-3aa76fbfb834 req-4283bb6c-227d-49cf-a75c-47a52d3fb886 service nova] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Received event network-vif-deleted-59301aa5-0ef4-43a3-87e6-1f968c41bd4f {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 756.318486] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ecf22e2-b73a-4dd8-aa25-6a7af5de4d44 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.326542] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f30d860f-2f66-4eec-86a3-85fdfb629a8f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.362826] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2675549f-a067-4cb8-a16b-7d9b545155e3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.370658] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749f4fe8-eef7-4b1f-aaed-986b8c421339 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.386438] env[61970]: DEBUG nova.compute.provider_tree [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.407400] env[61970]: DEBUG oslo_concurrency.lockutils [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Releasing lock "refresh_cache-fd4274ed-bf9a-45ab-a350-a8ebbc5ad404" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.407732] env[61970]: DEBUG nova.compute.manager [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 756.407977] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 756.408409] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f2bcc12-d844-42e9-b95d-3f0a129b00c6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.416863] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76744db8-a334-4a2a-a6c1-240caeba95bd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.438739] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fd4274ed-bf9a-45ab-a350-a8ebbc5ad404 could not be found. [ 756.439600] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 756.439600] env[61970]: INFO nova.compute.manager [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Took 0.03 seconds to destroy the instance on the hypervisor. [ 756.439600] env[61970]: DEBUG oslo.service.loopingcall [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 756.439766] env[61970]: DEBUG nova.compute.manager [-] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 756.439766] env[61970]: DEBUG nova.network.neutron [-] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 756.454064] env[61970]: DEBUG nova.network.neutron [-] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.602178] env[61970]: DEBUG nova.network.neutron [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.890701] env[61970]: DEBUG nova.scheduler.client.report [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 756.956473] env[61970]: DEBUG nova.network.neutron [-] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.104743] env[61970]: INFO nova.compute.manager [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] [instance: a540095e-563e-4059-b861-5d0e1b4995f4] Took 1.04 seconds to deallocate network for instance. [ 757.395797] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.439s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.396275] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 757.401751] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.700s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.461715] env[61970]: INFO nova.compute.manager [-] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Took 1.02 seconds to deallocate network for instance. [ 757.464779] env[61970]: DEBUG nova.compute.claims [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 757.464966] env[61970]: DEBUG oslo_concurrency.lockutils [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.910983] env[61970]: DEBUG nova.compute.utils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 757.912509] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 757.912705] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 757.970806] env[61970]: DEBUG nova.policy [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b07eb2474b1c4fbb98207a4d435500a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd9c85ee0fccc4373b4898f0256a69d38', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 758.143036] env[61970]: INFO nova.scheduler.client.report [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Deleted allocations for instance a540095e-563e-4059-b861-5d0e1b4995f4 [ 758.315379] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57027ed2-f33e-4dae-a350-d43df5ba62a8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.327356] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-323c03f7-0a3b-4fc9-b530-bb69db8160c5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.359128] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb0d5d0-7bf6-4446-99d3-1fe6f2467e5e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.367619] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0957eed-eb14-4012-8941-4fa8f2d7889e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.382080] env[61970]: DEBUG nova.compute.provider_tree [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.384069] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Successfully created port: 774d7fd3-a724-4311-a158-c727527f9cb3 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 758.416157] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 758.659388] env[61970]: DEBUG oslo_concurrency.lockutils [None req-86b40826-fce7-48ba-b13b-65e85e3e6d5d tempest-ImagesOneServerNegativeTestJSON-497980800 tempest-ImagesOneServerNegativeTestJSON-497980800-project-member] Lock "a540095e-563e-4059-b861-5d0e1b4995f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.156s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.887499] env[61970]: DEBUG nova.scheduler.client.report [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.162638] env[61970]: DEBUG nova.compute.manager [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 759.271766] env[61970]: DEBUG nova.compute.manager [req-38d3b2b0-8517-408d-8f62-471aea693246 req-2ad33262-5355-42dd-a3eb-04691352e8d9 service nova] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Received event network-changed-774d7fd3-a724-4311-a158-c727527f9cb3 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 759.272039] env[61970]: DEBUG nova.compute.manager [req-38d3b2b0-8517-408d-8f62-471aea693246 req-2ad33262-5355-42dd-a3eb-04691352e8d9 service nova] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Refreshing instance network info cache due to event network-changed-774d7fd3-a724-4311-a158-c727527f9cb3. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 759.272250] env[61970]: DEBUG oslo_concurrency.lockutils [req-38d3b2b0-8517-408d-8f62-471aea693246 req-2ad33262-5355-42dd-a3eb-04691352e8d9 service nova] Acquiring lock "refresh_cache-e74c9188-4f8a-4071-bc4f-c2be91fec3f7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.272659] env[61970]: DEBUG oslo_concurrency.lockutils [req-38d3b2b0-8517-408d-8f62-471aea693246 req-2ad33262-5355-42dd-a3eb-04691352e8d9 service nova] Acquired lock "refresh_cache-e74c9188-4f8a-4071-bc4f-c2be91fec3f7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.272659] env[61970]: DEBUG nova.network.neutron [req-38d3b2b0-8517-408d-8f62-471aea693246 req-2ad33262-5355-42dd-a3eb-04691352e8d9 service nova] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Refreshing network info cache for port 774d7fd3-a724-4311-a158-c727527f9cb3 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 759.397544] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.996s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.398202] env[61970]: ERROR nova.compute.manager [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dab2accb-1541-4bcf-8bde-d36049571418, please check neutron logs for more information. [ 759.398202] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Traceback (most recent call last): [ 759.398202] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 759.398202] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] self.driver.spawn(context, instance, image_meta, [ 759.398202] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 759.398202] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] self._vmops.spawn(context, instance, image_meta, injected_files, [ 759.398202] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 759.398202] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] vm_ref = self.build_virtual_machine(instance, [ 759.398202] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 759.398202] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] vif_infos = vmwarevif.get_vif_info(self._session, [ 759.398202] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 759.398582] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] for vif in network_info: [ 759.398582] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 759.398582] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] return self._sync_wrapper(fn, *args, **kwargs) [ 759.398582] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 759.398582] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] self.wait() [ 759.398582] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 759.398582] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] self[:] = self._gt.wait() [ 759.398582] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 759.398582] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] return self._exit_event.wait() [ 759.398582] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 759.398582] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] current.throw(*self._exc) [ 759.398582] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.398582] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] result = function(*args, **kwargs) [ 759.398955] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 759.398955] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] return func(*args, **kwargs) [ 759.398955] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 759.398955] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] raise e [ 759.398955] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 759.398955] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] nwinfo = self.network_api.allocate_for_instance( [ 759.398955] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.398955] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] created_port_ids = self._update_ports_for_instance( [ 759.398955] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.398955] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] with excutils.save_and_reraise_exception(): [ 759.398955] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.398955] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] self.force_reraise() [ 759.398955] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.399328] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] raise self.value [ 759.399328] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.399328] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] updated_port = self._update_port( [ 759.399328] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.399328] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] _ensure_no_port_binding_failure(port) [ 759.399328] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.399328] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] raise exception.PortBindingFailed(port_id=port['id']) [ 759.399328] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] nova.exception.PortBindingFailed: Binding failed for port dab2accb-1541-4bcf-8bde-d36049571418, please check neutron logs for more information. [ 759.399328] env[61970]: ERROR nova.compute.manager [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] [ 759.399328] env[61970]: DEBUG nova.compute.utils [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Binding failed for port dab2accb-1541-4bcf-8bde-d36049571418, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 759.400290] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.400s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.401120] env[61970]: DEBUG nova.objects.instance [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61970) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 759.403319] env[61970]: DEBUG nova.compute.manager [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Build of instance 29ac968a-37db-4c0d-9be9-9f577b53c533 was re-scheduled: Binding failed for port dab2accb-1541-4bcf-8bde-d36049571418, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 759.403759] env[61970]: DEBUG nova.compute.manager [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 759.403980] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "refresh_cache-29ac968a-37db-4c0d-9be9-9f577b53c533" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.408230] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "refresh_cache-29ac968a-37db-4c0d-9be9-9f577b53c533" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.408417] env[61970]: DEBUG nova.network.neutron [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 759.428265] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 759.467451] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 759.467681] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 759.467834] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 759.468079] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 759.468177] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 759.468339] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 759.468564] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 759.468723] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 759.468900] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 759.470156] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 759.470156] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 759.471470] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a024d25-3898-4986-9e90-7415ef006ea6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.475581] env[61970]: ERROR nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 774d7fd3-a724-4311-a158-c727527f9cb3, please check neutron logs for more information. [ 759.475581] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 759.475581] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 759.475581] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 759.475581] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.475581] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 759.475581] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.475581] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 759.475581] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.475581] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 759.475581] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.475581] env[61970]: ERROR nova.compute.manager raise self.value [ 759.475581] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.475581] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 759.475581] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.475581] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 759.476146] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.476146] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 759.476146] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 774d7fd3-a724-4311-a158-c727527f9cb3, please check neutron logs for more information. [ 759.476146] env[61970]: ERROR nova.compute.manager [ 759.476146] env[61970]: Traceback (most recent call last): [ 759.476146] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 759.476146] env[61970]: listener.cb(fileno) [ 759.476146] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.476146] env[61970]: result = function(*args, **kwargs) [ 759.476146] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 759.476146] env[61970]: return func(*args, **kwargs) [ 759.476146] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 759.476146] env[61970]: raise e [ 759.476146] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 759.476146] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 759.476146] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.476146] env[61970]: created_port_ids = self._update_ports_for_instance( [ 759.476146] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.476146] env[61970]: with excutils.save_and_reraise_exception(): [ 759.476146] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.476146] env[61970]: self.force_reraise() [ 759.476146] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.476146] env[61970]: raise self.value [ 759.476146] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.476146] env[61970]: updated_port = self._update_port( [ 759.476146] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.476146] env[61970]: _ensure_no_port_binding_failure(port) [ 759.476146] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.476146] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 759.477035] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 774d7fd3-a724-4311-a158-c727527f9cb3, please check neutron logs for more information. [ 759.477035] env[61970]: Removing descriptor: 15 [ 759.481718] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57c6d819-737c-443b-96c0-eee771088495 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.505460] env[61970]: ERROR nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 774d7fd3-a724-4311-a158-c727527f9cb3, please check neutron logs for more information. [ 759.505460] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Traceback (most recent call last): [ 759.505460] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 759.505460] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] yield resources [ 759.505460] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 759.505460] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] self.driver.spawn(context, instance, image_meta, [ 759.505460] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 759.505460] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 759.505460] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 759.505460] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] vm_ref = self.build_virtual_machine(instance, [ 759.505460] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 759.505931] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] vif_infos = vmwarevif.get_vif_info(self._session, [ 759.505931] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 759.505931] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] for vif in network_info: [ 759.505931] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 759.505931] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] return self._sync_wrapper(fn, *args, **kwargs) [ 759.505931] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 759.505931] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] self.wait() [ 759.505931] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 759.505931] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] self[:] = self._gt.wait() [ 759.505931] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 759.505931] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] return self._exit_event.wait() [ 759.505931] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 759.505931] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] current.throw(*self._exc) [ 759.506406] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.506406] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] result = function(*args, **kwargs) [ 759.506406] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 759.506406] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] return func(*args, **kwargs) [ 759.506406] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 759.506406] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] raise e [ 759.506406] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 759.506406] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] nwinfo = self.network_api.allocate_for_instance( [ 759.506406] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.506406] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] created_port_ids = self._update_ports_for_instance( [ 759.506406] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.506406] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] with excutils.save_and_reraise_exception(): [ 759.506406] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.506971] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] self.force_reraise() [ 759.506971] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.506971] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] raise self.value [ 759.506971] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.506971] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] updated_port = self._update_port( [ 759.506971] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.506971] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] _ensure_no_port_binding_failure(port) [ 759.506971] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.506971] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] raise exception.PortBindingFailed(port_id=port['id']) [ 759.506971] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] nova.exception.PortBindingFailed: Binding failed for port 774d7fd3-a724-4311-a158-c727527f9cb3, please check neutron logs for more information. [ 759.506971] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] [ 759.506971] env[61970]: INFO nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Terminating instance [ 759.508799] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "refresh_cache-e74c9188-4f8a-4071-bc4f-c2be91fec3f7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.689478] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.960668] env[61970]: DEBUG nova.network.neutron [req-38d3b2b0-8517-408d-8f62-471aea693246 req-2ad33262-5355-42dd-a3eb-04691352e8d9 service nova] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.971389] env[61970]: DEBUG nova.network.neutron [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.113202] env[61970]: DEBUG nova.network.neutron [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.118840] env[61970]: DEBUG nova.network.neutron [req-38d3b2b0-8517-408d-8f62-471aea693246 req-2ad33262-5355-42dd-a3eb-04691352e8d9 service nova] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.166262] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Acquiring lock "b385407b-1bdd-4c53-907c-cb4c8ce16cc7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.166544] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Lock "b385407b-1bdd-4c53-907c-cb4c8ce16cc7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.418195] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c1a7fa15-af66-46d9-a7df-8e8e59af2e4a tempest-ServersAdmin275Test-1057682006 tempest-ServersAdmin275Test-1057682006-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.420509] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 19.115s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.420509] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.420509] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61970) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 760.420509] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.045s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.429562] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c30f61f-9380-49f3-814f-d183fc1a336f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.435125] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d03e094-fe7f-463b-bf64-bb081b7055a6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.454544] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6acb0f-faf1-46cc-adaa-df4b04e87d57 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.462274] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-467a6110-8153-4b19-a0f9-498d3eb229d5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.500733] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181478MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61970) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 760.500895] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.615958] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "refresh_cache-29ac968a-37db-4c0d-9be9-9f577b53c533" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.616216] env[61970]: DEBUG nova.compute.manager [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 760.616399] env[61970]: DEBUG nova.compute.manager [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 760.616565] env[61970]: DEBUG nova.network.neutron [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 760.621025] env[61970]: DEBUG oslo_concurrency.lockutils [req-38d3b2b0-8517-408d-8f62-471aea693246 req-2ad33262-5355-42dd-a3eb-04691352e8d9 service nova] Releasing lock "refresh_cache-e74c9188-4f8a-4071-bc4f-c2be91fec3f7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.621445] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquired lock "refresh_cache-e74c9188-4f8a-4071-bc4f-c2be91fec3f7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.621609] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 760.633973] env[61970]: DEBUG nova.network.neutron [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.136013] env[61970]: DEBUG nova.network.neutron [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.145388] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.255572] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d139ad2-279d-481d-894c-fe462845b162 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.266217] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9977be94-a2b7-425b-9e50-0673f9cbb97d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.296089] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.297595] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d82a93be-aa49-43f7-90e1-e89a75b6348f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.305248] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c1de1aa-16d8-4d5e-a7b2-042d36818655 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.313319] env[61970]: DEBUG nova.compute.manager [req-b3ab99f2-2d66-4e30-8c80-eec191d188f4 req-dd0a66eb-8c19-4492-a55a-2f1ab7af374c service nova] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Received event network-vif-deleted-774d7fd3-a724-4311-a158-c727527f9cb3 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.323806] env[61970]: DEBUG nova.compute.provider_tree [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.382329] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Acquiring lock "6838c6a4-8241-4b6f-ad58-3acfd7afdbf4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.382638] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Lock "6838c6a4-8241-4b6f-ad58-3acfd7afdbf4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.644794] env[61970]: INFO nova.compute.manager [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 29ac968a-37db-4c0d-9be9-9f577b53c533] Took 1.03 seconds to deallocate network for instance. [ 761.801425] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Releasing lock "refresh_cache-e74c9188-4f8a-4071-bc4f-c2be91fec3f7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.802009] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 761.802271] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 761.802597] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9acba536-4700-4308-8999-40d58752a3f4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.811537] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2f17ad8-c5c7-4199-9d82-a87aad7e23f0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.826612] env[61970]: DEBUG nova.scheduler.client.report [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 761.834243] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e74c9188-4f8a-4071-bc4f-c2be91fec3f7 could not be found. [ 761.834243] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 761.834243] env[61970]: INFO nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Took 0.03 seconds to destroy the instance on the hypervisor. [ 761.834476] env[61970]: DEBUG oslo.service.loopingcall [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 761.834819] env[61970]: DEBUG nova.compute.manager [-] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 761.834916] env[61970]: DEBUG nova.network.neutron [-] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 761.855727] env[61970]: DEBUG nova.network.neutron [-] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.336870] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.916s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.337268] env[61970]: ERROR nova.compute.manager [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 663f1e4e-7249-473d-8538-7f4e66ba7d9c, please check neutron logs for more information. [ 762.337268] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Traceback (most recent call last): [ 762.337268] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 762.337268] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] self.driver.spawn(context, instance, image_meta, [ 762.337268] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 762.337268] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 762.337268] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 762.337268] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] vm_ref = self.build_virtual_machine(instance, [ 762.337268] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 762.337268] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] vif_infos = vmwarevif.get_vif_info(self._session, [ 762.337268] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 762.337609] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] for vif in network_info: [ 762.337609] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 762.337609] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] return self._sync_wrapper(fn, *args, **kwargs) [ 762.337609] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 762.337609] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] self.wait() [ 762.337609] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 762.337609] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] self[:] = self._gt.wait() [ 762.337609] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 762.337609] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] return self._exit_event.wait() [ 762.337609] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 762.337609] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] current.throw(*self._exc) [ 762.337609] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.337609] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] result = function(*args, **kwargs) [ 762.337975] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 762.337975] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] return func(*args, **kwargs) [ 762.337975] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 762.337975] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] raise e [ 762.337975] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.337975] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] nwinfo = self.network_api.allocate_for_instance( [ 762.337975] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.337975] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] created_port_ids = self._update_ports_for_instance( [ 762.337975] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.337975] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] with excutils.save_and_reraise_exception(): [ 762.337975] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.337975] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] self.force_reraise() [ 762.337975] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.338359] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] raise self.value [ 762.338359] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.338359] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] updated_port = self._update_port( [ 762.338359] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.338359] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] _ensure_no_port_binding_failure(port) [ 762.338359] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.338359] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] raise exception.PortBindingFailed(port_id=port['id']) [ 762.338359] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] nova.exception.PortBindingFailed: Binding failed for port 663f1e4e-7249-473d-8538-7f4e66ba7d9c, please check neutron logs for more information. [ 762.338359] env[61970]: ERROR nova.compute.manager [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] [ 762.338359] env[61970]: DEBUG nova.compute.utils [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Binding failed for port 663f1e4e-7249-473d-8538-7f4e66ba7d9c, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 762.338942] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.808s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.339181] env[61970]: DEBUG nova.objects.instance [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Lazy-loading 'resources' on Instance uuid 915c3b6c-640b-430f-b264-40bf85c642b9 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 762.340452] env[61970]: DEBUG nova.compute.manager [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Build of instance 34719fba-dc16-4b43-ab48-517982ca85a4 was re-scheduled: Binding failed for port 663f1e4e-7249-473d-8538-7f4e66ba7d9c, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 762.341045] env[61970]: DEBUG nova.compute.manager [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 762.341337] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Acquiring lock "refresh_cache-34719fba-dc16-4b43-ab48-517982ca85a4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.341418] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Acquired lock "refresh_cache-34719fba-dc16-4b43-ab48-517982ca85a4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.341596] env[61970]: DEBUG nova.network.neutron [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 762.358118] env[61970]: DEBUG nova.network.neutron [-] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.681971] env[61970]: INFO nova.scheduler.client.report [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleted allocations for instance 29ac968a-37db-4c0d-9be9-9f577b53c533 [ 762.859806] env[61970]: DEBUG nova.network.neutron [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.861746] env[61970]: INFO nova.compute.manager [-] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Took 1.03 seconds to deallocate network for instance. [ 762.866355] env[61970]: DEBUG nova.compute.claims [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 762.867031] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.964477] env[61970]: DEBUG nova.network.neutron [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.126520] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0dd611-b2ba-4b9a-9226-dccfb9ceb5f5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.133717] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac91146d-d118-48e4-a2f2-c8b5fd1c893e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.163508] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c2750c-9df8-4dea-933a-fd3e27f644d5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.170589] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b0713bf-3370-4d1d-965f-f85eae31d776 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.183323] env[61970]: DEBUG nova.compute.provider_tree [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.192209] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d400307b-1e59-426a-8ece-29f891788b9d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "29ac968a-37db-4c0d-9be9-9f577b53c533" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.549s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.466180] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Releasing lock "refresh_cache-34719fba-dc16-4b43-ab48-517982ca85a4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.466455] env[61970]: DEBUG nova.compute.manager [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 763.467025] env[61970]: DEBUG nova.compute.manager [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 763.467025] env[61970]: DEBUG nova.network.neutron [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 763.485595] env[61970]: DEBUG nova.network.neutron [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.686299] env[61970]: DEBUG nova.scheduler.client.report [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 763.694847] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 763.988252] env[61970]: DEBUG nova.network.neutron [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.192120] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.853s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.194608] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.583s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.214032] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.220180] env[61970]: INFO nova.scheduler.client.report [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Deleted allocations for instance 915c3b6c-640b-430f-b264-40bf85c642b9 [ 764.491232] env[61970]: INFO nova.compute.manager [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] [instance: 34719fba-dc16-4b43-ab48-517982ca85a4] Took 1.02 seconds to deallocate network for instance. [ 764.731396] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e1f9b21a-6a55-4838-bd16-139036b0adb0 tempest-ServersAdmin275Test-2076836576 tempest-ServersAdmin275Test-2076836576-project-member] Lock "915c3b6c-640b-430f-b264-40bf85c642b9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.683s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.953650] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "d09dbde2-5daa-4ae9-9fd5-bfbb03fba136" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.954930] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "d09dbde2-5daa-4ae9-9fd5-bfbb03fba136" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.027236] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01261fe1-fa34-4af9-978b-a75797afaec9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.035125] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0541c24-80e4-4c48-9cce-8a16c7e8c298 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.066010] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00a15a5d-691c-4367-9d6e-3bc5a50e1f06 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.073259] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f72c09-a50e-4e83-a870-17b02a934a50 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.086283] env[61970]: DEBUG nova.compute.provider_tree [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.521625] env[61970]: INFO nova.scheduler.client.report [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Deleted allocations for instance 34719fba-dc16-4b43-ab48-517982ca85a4 [ 765.590453] env[61970]: DEBUG nova.scheduler.client.report [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 766.032384] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b6260520-a456-418d-b38e-707bda514453 tempest-ServerMetadataNegativeTestJSON-731742410 tempest-ServerMetadataNegativeTestJSON-731742410-project-member] Lock "34719fba-dc16-4b43-ab48-517982ca85a4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.435s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.094837] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.900s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.095474] env[61970]: ERROR nova.compute.manager [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 242ccfc0-6f0b-43ae-b607-a09016baa251, please check neutron logs for more information. [ 766.095474] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Traceback (most recent call last): [ 766.095474] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 766.095474] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] self.driver.spawn(context, instance, image_meta, [ 766.095474] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 766.095474] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] self._vmops.spawn(context, instance, image_meta, injected_files, [ 766.095474] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 766.095474] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] vm_ref = self.build_virtual_machine(instance, [ 766.095474] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 766.095474] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] vif_infos = vmwarevif.get_vif_info(self._session, [ 766.095474] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 766.095936] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] for vif in network_info: [ 766.095936] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 766.095936] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] return self._sync_wrapper(fn, *args, **kwargs) [ 766.095936] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 766.095936] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] self.wait() [ 766.095936] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 766.095936] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] self[:] = self._gt.wait() [ 766.095936] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 766.095936] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] return self._exit_event.wait() [ 766.095936] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 766.095936] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] result = hub.switch() [ 766.095936] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 766.095936] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] return self.greenlet.switch() [ 766.096371] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.096371] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] result = function(*args, **kwargs) [ 766.096371] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 766.096371] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] return func(*args, **kwargs) [ 766.096371] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 766.096371] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] raise e [ 766.096371] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.096371] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] nwinfo = self.network_api.allocate_for_instance( [ 766.096371] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.096371] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] created_port_ids = self._update_ports_for_instance( [ 766.096371] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.096371] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] with excutils.save_and_reraise_exception(): [ 766.096371] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.096788] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] self.force_reraise() [ 766.096788] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.096788] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] raise self.value [ 766.096788] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.096788] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] updated_port = self._update_port( [ 766.096788] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.096788] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] _ensure_no_port_binding_failure(port) [ 766.096788] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.096788] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] raise exception.PortBindingFailed(port_id=port['id']) [ 766.096788] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] nova.exception.PortBindingFailed: Binding failed for port 242ccfc0-6f0b-43ae-b607-a09016baa251, please check neutron logs for more information. [ 766.096788] env[61970]: ERROR nova.compute.manager [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] [ 766.097231] env[61970]: DEBUG nova.compute.utils [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Binding failed for port 242ccfc0-6f0b-43ae-b607-a09016baa251, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 766.097427] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.356s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.100216] env[61970]: DEBUG nova.compute.manager [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Build of instance 1dc83f19-2965-418b-8d2a-56aa1124d258 was re-scheduled: Binding failed for port 242ccfc0-6f0b-43ae-b607-a09016baa251, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 766.100671] env[61970]: DEBUG nova.compute.manager [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 766.100900] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Acquiring lock "refresh_cache-1dc83f19-2965-418b-8d2a-56aa1124d258" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.101237] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Acquired lock "refresh_cache-1dc83f19-2965-418b-8d2a-56aa1124d258" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.101237] env[61970]: DEBUG nova.network.neutron [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 766.317808] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Acquiring lock "e1415314-f137-4ee3-a065-3d875ef9e7ff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.318035] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Lock "e1415314-f137-4ee3-a065-3d875ef9e7ff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.535451] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 766.623058] env[61970]: DEBUG nova.network.neutron [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.738139] env[61970]: DEBUG nova.network.neutron [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.900812] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db7dbbff-87b7-4d40-b4a2-eb7aa56e4f31 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.908482] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-447e3506-7a85-4787-b3f3-5aa746187855 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.937043] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a89a537-9b28-4960-9f18-e32819efb710 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.944696] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de2146c9-f04c-48f2-9dfb-5998ae327475 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.961022] env[61970]: DEBUG nova.compute.provider_tree [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.056012] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.240602] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Releasing lock "refresh_cache-1dc83f19-2965-418b-8d2a-56aa1124d258" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.240876] env[61970]: DEBUG nova.compute.manager [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 767.241070] env[61970]: DEBUG nova.compute.manager [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 767.241241] env[61970]: DEBUG nova.network.neutron [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 767.258871] env[61970]: DEBUG nova.network.neutron [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.461463] env[61970]: DEBUG nova.scheduler.client.report [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 767.762537] env[61970]: DEBUG nova.network.neutron [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.966786] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.869s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.967512] env[61970]: ERROR nova.compute.manager [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4026ce5e-c59e-459c-b80b-1ee765d383bf, please check neutron logs for more information. [ 767.967512] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Traceback (most recent call last): [ 767.967512] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 767.967512] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] self.driver.spawn(context, instance, image_meta, [ 767.967512] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 767.967512] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] self._vmops.spawn(context, instance, image_meta, injected_files, [ 767.967512] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 767.967512] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] vm_ref = self.build_virtual_machine(instance, [ 767.967512] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 767.967512] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] vif_infos = vmwarevif.get_vif_info(self._session, [ 767.967512] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 767.967930] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] for vif in network_info: [ 767.967930] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 767.967930] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] return self._sync_wrapper(fn, *args, **kwargs) [ 767.967930] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 767.967930] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] self.wait() [ 767.967930] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 767.967930] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] self[:] = self._gt.wait() [ 767.967930] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 767.967930] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] return self._exit_event.wait() [ 767.967930] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 767.967930] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] result = hub.switch() [ 767.967930] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 767.967930] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] return self.greenlet.switch() [ 767.968396] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.968396] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] result = function(*args, **kwargs) [ 767.968396] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 767.968396] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] return func(*args, **kwargs) [ 767.968396] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 767.968396] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] raise e [ 767.968396] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.968396] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] nwinfo = self.network_api.allocate_for_instance( [ 767.968396] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.968396] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] created_port_ids = self._update_ports_for_instance( [ 767.968396] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.968396] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] with excutils.save_and_reraise_exception(): [ 767.968396] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.968815] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] self.force_reraise() [ 767.968815] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.968815] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] raise self.value [ 767.968815] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.968815] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] updated_port = self._update_port( [ 767.968815] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.968815] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] _ensure_no_port_binding_failure(port) [ 767.968815] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.968815] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] raise exception.PortBindingFailed(port_id=port['id']) [ 767.968815] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] nova.exception.PortBindingFailed: Binding failed for port 4026ce5e-c59e-459c-b80b-1ee765d383bf, please check neutron logs for more information. [ 767.968815] env[61970]: ERROR nova.compute.manager [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] [ 767.969252] env[61970]: DEBUG nova.compute.utils [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Binding failed for port 4026ce5e-c59e-459c-b80b-1ee765d383bf, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 767.969474] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.058s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.970902] env[61970]: INFO nova.compute.claims [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 767.974041] env[61970]: DEBUG nova.compute.manager [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Build of instance ed10e0a6-c79f-4783-bb93-312c5f0ffd51 was re-scheduled: Binding failed for port 4026ce5e-c59e-459c-b80b-1ee765d383bf, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 767.974588] env[61970]: DEBUG nova.compute.manager [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 767.974815] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Acquiring lock "refresh_cache-ed10e0a6-c79f-4783-bb93-312c5f0ffd51" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.974963] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Acquired lock "refresh_cache-ed10e0a6-c79f-4783-bb93-312c5f0ffd51" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.975134] env[61970]: DEBUG nova.network.neutron [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 768.265303] env[61970]: INFO nova.compute.manager [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] [instance: 1dc83f19-2965-418b-8d2a-56aa1124d258] Took 1.02 seconds to deallocate network for instance. [ 768.498975] env[61970]: DEBUG nova.network.neutron [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.586096] env[61970]: DEBUG nova.network.neutron [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.088546] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Releasing lock "refresh_cache-ed10e0a6-c79f-4783-bb93-312c5f0ffd51" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.088830] env[61970]: DEBUG nova.compute.manager [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 769.088978] env[61970]: DEBUG nova.compute.manager [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 769.089166] env[61970]: DEBUG nova.network.neutron [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 769.104143] env[61970]: DEBUG nova.network.neutron [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.260649] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a120bf2-a7c4-41c1-8d4f-3acde5c1797d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.267906] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b59a24c3-dc0b-43a5-92e7-1008a242fa91 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.299792] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24c478fb-1ff9-4172-a8ee-921e518da165 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.306702] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db41204-a303-4650-9853-94cee781681c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.319632] env[61970]: DEBUG nova.compute.provider_tree [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.321635] env[61970]: INFO nova.scheduler.client.report [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Deleted allocations for instance 1dc83f19-2965-418b-8d2a-56aa1124d258 [ 769.607367] env[61970]: DEBUG nova.network.neutron [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.828573] env[61970]: DEBUG nova.scheduler.client.report [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 769.831804] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6777e88a-3ad8-4158-bc53-5bebe5d01aee tempest-ServersTestMultiNic-323107048 tempest-ServersTestMultiNic-323107048-project-member] Lock "1dc83f19-2965-418b-8d2a-56aa1124d258" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.111s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.112409] env[61970]: INFO nova.compute.manager [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: ed10e0a6-c79f-4783-bb93-312c5f0ffd51] Took 1.02 seconds to deallocate network for instance. [ 770.333526] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.364s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.334113] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 770.336680] env[61970]: DEBUG oslo_concurrency.lockutils [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.725s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.339696] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 770.845850] env[61970]: DEBUG nova.compute.utils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 770.849178] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 770.850025] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 770.867010] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.908380] env[61970]: DEBUG nova.policy [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b07eb2474b1c4fbb98207a4d435500a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd9c85ee0fccc4373b4898f0256a69d38', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 771.153256] env[61970]: INFO nova.scheduler.client.report [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Deleted allocations for instance ed10e0a6-c79f-4783-bb93-312c5f0ffd51 [ 771.181274] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac19a5c4-10ce-4874-880a-efd3f3f7d217 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.188462] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472710a4-378f-4b16-9a7c-66316ebbc001 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.222667] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad62daf8-9ed6-40b0-a428-470b0f48da83 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.230843] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-135e7ff9-4757-4aed-92ef-a6899c35436f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.247061] env[61970]: DEBUG nova.compute.provider_tree [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.354661] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 771.580067] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Successfully created port: e37d7d20-8753-4ad6-98b7-c7dee3b1c724 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 771.665274] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2b54935e-2f53-4431-990d-5e7779ac52e6 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Lock "ed10e0a6-c79f-4783-bb93-312c5f0ffd51" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.866s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.750587] env[61970]: DEBUG nova.scheduler.client.report [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 772.169317] env[61970]: DEBUG nova.compute.manager [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 772.259136] env[61970]: DEBUG oslo_concurrency.lockutils [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.922s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.260107] env[61970]: ERROR nova.compute.manager [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 35fe4568-400f-49b0-8ad9-0925de026db3, please check neutron logs for more information. [ 772.260107] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Traceback (most recent call last): [ 772.260107] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 772.260107] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] self.driver.spawn(context, instance, image_meta, [ 772.260107] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 772.260107] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 772.260107] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 772.260107] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] vm_ref = self.build_virtual_machine(instance, [ 772.260107] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 772.260107] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] vif_infos = vmwarevif.get_vif_info(self._session, [ 772.260107] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 772.260561] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] for vif in network_info: [ 772.260561] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 772.260561] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] return self._sync_wrapper(fn, *args, **kwargs) [ 772.260561] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 772.260561] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] self.wait() [ 772.260561] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 772.260561] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] self[:] = self._gt.wait() [ 772.260561] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 772.260561] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] return self._exit_event.wait() [ 772.260561] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 772.260561] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] current.throw(*self._exc) [ 772.260561] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 772.260561] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] result = function(*args, **kwargs) [ 772.260956] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 772.260956] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] return func(*args, **kwargs) [ 772.260956] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 772.260956] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] raise e [ 772.260956] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 772.260956] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] nwinfo = self.network_api.allocate_for_instance( [ 772.260956] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 772.260956] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] created_port_ids = self._update_ports_for_instance( [ 772.260956] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 772.260956] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] with excutils.save_and_reraise_exception(): [ 772.260956] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.260956] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] self.force_reraise() [ 772.260956] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.261336] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] raise self.value [ 772.261336] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 772.261336] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] updated_port = self._update_port( [ 772.261336] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.261336] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] _ensure_no_port_binding_failure(port) [ 772.261336] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.261336] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] raise exception.PortBindingFailed(port_id=port['id']) [ 772.261336] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] nova.exception.PortBindingFailed: Binding failed for port 35fe4568-400f-49b0-8ad9-0925de026db3, please check neutron logs for more information. [ 772.261336] env[61970]: ERROR nova.compute.manager [instance: 5d352f39-4544-46ec-af43-e539d511b47b] [ 772.261336] env[61970]: DEBUG nova.compute.utils [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Binding failed for port 35fe4568-400f-49b0-8ad9-0925de026db3, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 772.261670] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.012s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.265085] env[61970]: INFO nova.compute.claims [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 772.265657] env[61970]: DEBUG nova.compute.manager [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Build of instance 5d352f39-4544-46ec-af43-e539d511b47b was re-scheduled: Binding failed for port 35fe4568-400f-49b0-8ad9-0925de026db3, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 772.266141] env[61970]: DEBUG nova.compute.manager [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 772.266364] env[61970]: DEBUG oslo_concurrency.lockutils [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Acquiring lock "refresh_cache-5d352f39-4544-46ec-af43-e539d511b47b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.266503] env[61970]: DEBUG oslo_concurrency.lockutils [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Acquired lock "refresh_cache-5d352f39-4544-46ec-af43-e539d511b47b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.266651] env[61970]: DEBUG nova.network.neutron [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 772.362924] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 772.391790] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 772.392067] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 772.392236] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 772.392416] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 772.392561] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 772.392706] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 772.392909] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 772.393210] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 772.393408] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 772.393578] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 772.393751] env[61970]: DEBUG nova.virt.hardware [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 772.394607] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a8f4b69-2c45-4496-abcd-8eff1750ffa7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.402699] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642b2e38-3dc9-4ad5-b682-98f90bfbcc7b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.690503] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.763433] env[61970]: DEBUG nova.compute.manager [req-1d5a8d38-81e2-44c0-a7e9-b04e352616ef req-ac9884bf-f3ca-4f2a-8a98-2b576a9f1517 service nova] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Received event network-changed-e37d7d20-8753-4ad6-98b7-c7dee3b1c724 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 772.763433] env[61970]: DEBUG nova.compute.manager [req-1d5a8d38-81e2-44c0-a7e9-b04e352616ef req-ac9884bf-f3ca-4f2a-8a98-2b576a9f1517 service nova] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Refreshing instance network info cache due to event network-changed-e37d7d20-8753-4ad6-98b7-c7dee3b1c724. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 772.763433] env[61970]: DEBUG oslo_concurrency.lockutils [req-1d5a8d38-81e2-44c0-a7e9-b04e352616ef req-ac9884bf-f3ca-4f2a-8a98-2b576a9f1517 service nova] Acquiring lock "refresh_cache-01ee68e2-4310-477e-8adb-6ad67d6a65e7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.763693] env[61970]: DEBUG oslo_concurrency.lockutils [req-1d5a8d38-81e2-44c0-a7e9-b04e352616ef req-ac9884bf-f3ca-4f2a-8a98-2b576a9f1517 service nova] Acquired lock "refresh_cache-01ee68e2-4310-477e-8adb-6ad67d6a65e7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.763693] env[61970]: DEBUG nova.network.neutron [req-1d5a8d38-81e2-44c0-a7e9-b04e352616ef req-ac9884bf-f3ca-4f2a-8a98-2b576a9f1517 service nova] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Refreshing network info cache for port e37d7d20-8753-4ad6-98b7-c7dee3b1c724 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 772.804721] env[61970]: DEBUG nova.network.neutron [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.933134] env[61970]: DEBUG nova.network.neutron [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.976709] env[61970]: ERROR nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e37d7d20-8753-4ad6-98b7-c7dee3b1c724, please check neutron logs for more information. [ 772.976709] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 772.976709] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 772.976709] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 772.976709] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 772.976709] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 772.976709] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 772.976709] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 772.976709] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.976709] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 772.976709] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.976709] env[61970]: ERROR nova.compute.manager raise self.value [ 772.976709] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 772.976709] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 772.976709] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.976709] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 772.977251] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.977251] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 772.977251] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e37d7d20-8753-4ad6-98b7-c7dee3b1c724, please check neutron logs for more information. [ 772.977251] env[61970]: ERROR nova.compute.manager [ 772.977251] env[61970]: Traceback (most recent call last): [ 772.977251] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 772.977251] env[61970]: listener.cb(fileno) [ 772.977251] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 772.977251] env[61970]: result = function(*args, **kwargs) [ 772.977251] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 772.977251] env[61970]: return func(*args, **kwargs) [ 772.977251] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 772.977251] env[61970]: raise e [ 772.977251] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 772.977251] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 772.977251] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 772.977251] env[61970]: created_port_ids = self._update_ports_for_instance( [ 772.977251] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 772.977251] env[61970]: with excutils.save_and_reraise_exception(): [ 772.977251] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.977251] env[61970]: self.force_reraise() [ 772.977251] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.977251] env[61970]: raise self.value [ 772.977251] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 772.977251] env[61970]: updated_port = self._update_port( [ 772.977251] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.977251] env[61970]: _ensure_no_port_binding_failure(port) [ 772.977251] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.977251] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 772.978194] env[61970]: nova.exception.PortBindingFailed: Binding failed for port e37d7d20-8753-4ad6-98b7-c7dee3b1c724, please check neutron logs for more information. [ 772.978194] env[61970]: Removing descriptor: 15 [ 772.978194] env[61970]: ERROR nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e37d7d20-8753-4ad6-98b7-c7dee3b1c724, please check neutron logs for more information. [ 772.978194] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Traceback (most recent call last): [ 772.978194] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 772.978194] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] yield resources [ 772.978194] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 772.978194] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] self.driver.spawn(context, instance, image_meta, [ 772.978194] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 772.978194] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 772.978194] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 772.978194] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] vm_ref = self.build_virtual_machine(instance, [ 772.978633] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 772.978633] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] vif_infos = vmwarevif.get_vif_info(self._session, [ 772.978633] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 772.978633] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] for vif in network_info: [ 772.978633] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 772.978633] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] return self._sync_wrapper(fn, *args, **kwargs) [ 772.978633] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 772.978633] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] self.wait() [ 772.978633] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 772.978633] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] self[:] = self._gt.wait() [ 772.978633] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 772.978633] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] return self._exit_event.wait() [ 772.978633] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 772.979159] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] result = hub.switch() [ 772.979159] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 772.979159] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] return self.greenlet.switch() [ 772.979159] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 772.979159] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] result = function(*args, **kwargs) [ 772.979159] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 772.979159] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] return func(*args, **kwargs) [ 772.979159] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 772.979159] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] raise e [ 772.979159] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 772.979159] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] nwinfo = self.network_api.allocate_for_instance( [ 772.979159] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 772.979159] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] created_port_ids = self._update_ports_for_instance( [ 772.979616] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 772.979616] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] with excutils.save_and_reraise_exception(): [ 772.979616] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.979616] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] self.force_reraise() [ 772.979616] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.979616] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] raise self.value [ 772.979616] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 772.979616] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] updated_port = self._update_port( [ 772.979616] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.979616] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] _ensure_no_port_binding_failure(port) [ 772.979616] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.979616] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] raise exception.PortBindingFailed(port_id=port['id']) [ 772.979973] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] nova.exception.PortBindingFailed: Binding failed for port e37d7d20-8753-4ad6-98b7-c7dee3b1c724, please check neutron logs for more information. [ 772.979973] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] [ 772.979973] env[61970]: INFO nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Terminating instance [ 772.979973] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "refresh_cache-01ee68e2-4310-477e-8adb-6ad67d6a65e7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.286659] env[61970]: DEBUG nova.network.neutron [req-1d5a8d38-81e2-44c0-a7e9-b04e352616ef req-ac9884bf-f3ca-4f2a-8a98-2b576a9f1517 service nova] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.365241] env[61970]: DEBUG nova.network.neutron [req-1d5a8d38-81e2-44c0-a7e9-b04e352616ef req-ac9884bf-f3ca-4f2a-8a98-2b576a9f1517 service nova] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.436021] env[61970]: DEBUG oslo_concurrency.lockutils [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Releasing lock "refresh_cache-5d352f39-4544-46ec-af43-e539d511b47b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.436021] env[61970]: DEBUG nova.compute.manager [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 773.436287] env[61970]: DEBUG nova.compute.manager [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 773.436356] env[61970]: DEBUG nova.network.neutron [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 773.453131] env[61970]: DEBUG nova.network.neutron [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.550692] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a9e591-ee5b-4f30-9530-760f637421f6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.558313] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f67d2fd-5955-4b26-86f9-9904608beab6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.590201] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8948df3-5f6d-4697-9f40-63a0911104c6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.598274] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c7ba15-ba22-4468-84e6-9d74aa36477a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.612064] env[61970]: DEBUG nova.compute.provider_tree [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.867554] env[61970]: DEBUG oslo_concurrency.lockutils [req-1d5a8d38-81e2-44c0-a7e9-b04e352616ef req-ac9884bf-f3ca-4f2a-8a98-2b576a9f1517 service nova] Releasing lock "refresh_cache-01ee68e2-4310-477e-8adb-6ad67d6a65e7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.868066] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquired lock "refresh_cache-01ee68e2-4310-477e-8adb-6ad67d6a65e7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.868208] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 773.956031] env[61970]: DEBUG nova.network.neutron [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.115264] env[61970]: DEBUG nova.scheduler.client.report [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.390085] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.459053] env[61970]: INFO nova.compute.manager [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] [instance: 5d352f39-4544-46ec-af43-e539d511b47b] Took 1.02 seconds to deallocate network for instance. [ 774.464930] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.620672] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.359s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.621285] env[61970]: DEBUG nova.compute.manager [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 774.623828] env[61970]: DEBUG oslo_concurrency.lockutils [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.159s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.792310] env[61970]: DEBUG nova.compute.manager [req-dba26da8-039a-43fe-abb8-29969650c752 req-436b5c2d-d42f-42a0-8ab5-1ebc23a3e06b service nova] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Received event network-vif-deleted-e37d7d20-8753-4ad6-98b7-c7dee3b1c724 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 774.968764] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Releasing lock "refresh_cache-01ee68e2-4310-477e-8adb-6ad67d6a65e7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.968897] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 774.969755] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 774.969755] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9fde7b1a-9c95-41ff-9a38-6ac3e144fcef {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.979693] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-550546f9-c620-49c2-a949-9c66b079baf9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.007189] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 01ee68e2-4310-477e-8adb-6ad67d6a65e7 could not be found. [ 775.007434] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 775.007612] env[61970]: INFO nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 775.007850] env[61970]: DEBUG oslo.service.loopingcall [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 775.008074] env[61970]: DEBUG nova.compute.manager [-] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 775.008171] env[61970]: DEBUG nova.network.neutron [-] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 775.022749] env[61970]: DEBUG nova.network.neutron [-] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 775.130707] env[61970]: DEBUG nova.compute.utils [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 775.137011] env[61970]: DEBUG nova.compute.manager [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 775.137011] env[61970]: DEBUG nova.network.neutron [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 775.177091] env[61970]: DEBUG nova.policy [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8b9f034f7f474448bfd555c4e8f88f48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '375eb5f90b7141e5a34ab9d79b7f93d5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 775.482622] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9f2ae6-4c9c-4d73-80c8-456fd6283878 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.487269] env[61970]: DEBUG nova.network.neutron [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Successfully created port: 5b0c073f-e9de-4f25-ab23-376446d9aad6 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 775.494472] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c44de1e-e5f0-4d9c-9c0d-2c6f714e6317 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.498511] env[61970]: INFO nova.scheduler.client.report [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Deleted allocations for instance 5d352f39-4544-46ec-af43-e539d511b47b [ 775.534610] env[61970]: DEBUG nova.network.neutron [-] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.537183] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b4c27f-04aa-4b47-a2b7-38241c945625 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.547028] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82df93b5-2a26-417b-8cf2-5e65e5665538 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.561280] env[61970]: DEBUG nova.compute.provider_tree [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.636159] env[61970]: DEBUG nova.compute.manager [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 776.008347] env[61970]: DEBUG oslo_concurrency.lockutils [None req-677e646e-e53e-44c9-89af-47400df2d643 tempest-ServersAdminTestJSON-1668139568 tempest-ServersAdminTestJSON-1668139568-project-member] Lock "5d352f39-4544-46ec-af43-e539d511b47b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.871s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.037589] env[61970]: INFO nova.compute.manager [-] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Took 1.03 seconds to deallocate network for instance. [ 776.039774] env[61970]: DEBUG nova.compute.claims [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 776.040291] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.063754] env[61970]: DEBUG nova.scheduler.client.report [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.512988] env[61970]: DEBUG nova.compute.manager [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 776.530495] env[61970]: ERROR nova.compute.manager [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5b0c073f-e9de-4f25-ab23-376446d9aad6, please check neutron logs for more information. [ 776.530495] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 776.530495] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.530495] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 776.530495] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 776.530495] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 776.530495] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 776.530495] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 776.530495] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.530495] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 776.530495] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.530495] env[61970]: ERROR nova.compute.manager raise self.value [ 776.530495] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 776.530495] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 776.530495] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.530495] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 776.531068] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.531068] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 776.531068] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5b0c073f-e9de-4f25-ab23-376446d9aad6, please check neutron logs for more information. [ 776.531068] env[61970]: ERROR nova.compute.manager [ 776.531068] env[61970]: Traceback (most recent call last): [ 776.531068] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 776.531068] env[61970]: listener.cb(fileno) [ 776.531068] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 776.531068] env[61970]: result = function(*args, **kwargs) [ 776.531068] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 776.531068] env[61970]: return func(*args, **kwargs) [ 776.531068] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 776.531068] env[61970]: raise e [ 776.531068] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.531068] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 776.531068] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 776.531068] env[61970]: created_port_ids = self._update_ports_for_instance( [ 776.531068] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 776.531068] env[61970]: with excutils.save_and_reraise_exception(): [ 776.531068] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.531068] env[61970]: self.force_reraise() [ 776.531068] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.531068] env[61970]: raise self.value [ 776.531068] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 776.531068] env[61970]: updated_port = self._update_port( [ 776.531068] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.531068] env[61970]: _ensure_no_port_binding_failure(port) [ 776.531068] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.531068] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 776.532087] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 5b0c073f-e9de-4f25-ab23-376446d9aad6, please check neutron logs for more information. [ 776.532087] env[61970]: Removing descriptor: 15 [ 776.568258] env[61970]: DEBUG oslo_concurrency.lockutils [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.944s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.568886] env[61970]: ERROR nova.compute.manager [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 59301aa5-0ef4-43a3-87e6-1f968c41bd4f, please check neutron logs for more information. [ 776.568886] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Traceback (most recent call last): [ 776.568886] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 776.568886] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] self.driver.spawn(context, instance, image_meta, [ 776.568886] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 776.568886] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] self._vmops.spawn(context, instance, image_meta, injected_files, [ 776.568886] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 776.568886] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] vm_ref = self.build_virtual_machine(instance, [ 776.568886] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 776.568886] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] vif_infos = vmwarevif.get_vif_info(self._session, [ 776.568886] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 776.569291] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] for vif in network_info: [ 776.569291] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 776.569291] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] return self._sync_wrapper(fn, *args, **kwargs) [ 776.569291] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 776.569291] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] self.wait() [ 776.569291] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 776.569291] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] self[:] = self._gt.wait() [ 776.569291] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 776.569291] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] return self._exit_event.wait() [ 776.569291] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 776.569291] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] result = hub.switch() [ 776.569291] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 776.569291] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] return self.greenlet.switch() [ 776.569711] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 776.569711] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] result = function(*args, **kwargs) [ 776.569711] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 776.569711] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] return func(*args, **kwargs) [ 776.569711] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 776.569711] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] raise e [ 776.569711] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.569711] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] nwinfo = self.network_api.allocate_for_instance( [ 776.569711] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 776.569711] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] created_port_ids = self._update_ports_for_instance( [ 776.569711] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 776.569711] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] with excutils.save_and_reraise_exception(): [ 776.569711] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.570148] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] self.force_reraise() [ 776.570148] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.570148] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] raise self.value [ 776.570148] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 776.570148] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] updated_port = self._update_port( [ 776.570148] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.570148] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] _ensure_no_port_binding_failure(port) [ 776.570148] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.570148] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] raise exception.PortBindingFailed(port_id=port['id']) [ 776.570148] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] nova.exception.PortBindingFailed: Binding failed for port 59301aa5-0ef4-43a3-87e6-1f968c41bd4f, please check neutron logs for more information. [ 776.570148] env[61970]: ERROR nova.compute.manager [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] [ 776.570515] env[61970]: DEBUG nova.compute.utils [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Binding failed for port 59301aa5-0ef4-43a3-87e6-1f968c41bd4f, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 776.571196] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.882s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.572921] env[61970]: INFO nova.compute.claims [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 776.576989] env[61970]: DEBUG nova.compute.manager [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Build of instance fd4274ed-bf9a-45ab-a350-a8ebbc5ad404 was re-scheduled: Binding failed for port 59301aa5-0ef4-43a3-87e6-1f968c41bd4f, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 776.577571] env[61970]: DEBUG nova.compute.manager [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 776.577931] env[61970]: DEBUG oslo_concurrency.lockutils [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "refresh_cache-fd4274ed-bf9a-45ab-a350-a8ebbc5ad404" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.578548] env[61970]: DEBUG oslo_concurrency.lockutils [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquired lock "refresh_cache-fd4274ed-bf9a-45ab-a350-a8ebbc5ad404" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.578792] env[61970]: DEBUG nova.network.neutron [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 776.645242] env[61970]: DEBUG nova.compute.manager [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 776.670181] env[61970]: DEBUG nova.virt.hardware [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 776.671080] env[61970]: DEBUG nova.virt.hardware [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 776.671080] env[61970]: DEBUG nova.virt.hardware [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 776.671080] env[61970]: DEBUG nova.virt.hardware [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 776.671080] env[61970]: DEBUG nova.virt.hardware [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 776.671080] env[61970]: DEBUG nova.virt.hardware [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 776.671315] env[61970]: DEBUG nova.virt.hardware [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 776.671473] env[61970]: DEBUG nova.virt.hardware [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 776.671637] env[61970]: DEBUG nova.virt.hardware [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 776.671791] env[61970]: DEBUG nova.virt.hardware [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 776.671956] env[61970]: DEBUG nova.virt.hardware [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 776.672836] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48949667-2b9f-4fa2-92d1-e943afe21846 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.681679] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7bdc5be-db26-4099-a0c6-17f95c51e74c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.695549] env[61970]: ERROR nova.compute.manager [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5b0c073f-e9de-4f25-ab23-376446d9aad6, please check neutron logs for more information. [ 776.695549] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] Traceback (most recent call last): [ 776.695549] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 776.695549] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] yield resources [ 776.695549] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 776.695549] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] self.driver.spawn(context, instance, image_meta, [ 776.695549] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 776.695549] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] self._vmops.spawn(context, instance, image_meta, injected_files, [ 776.695549] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 776.695549] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] vm_ref = self.build_virtual_machine(instance, [ 776.695549] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 776.696759] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] vif_infos = vmwarevif.get_vif_info(self._session, [ 776.696759] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 776.696759] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] for vif in network_info: [ 776.696759] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 776.696759] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] return self._sync_wrapper(fn, *args, **kwargs) [ 776.696759] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 776.696759] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] self.wait() [ 776.696759] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 776.696759] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] self[:] = self._gt.wait() [ 776.696759] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 776.696759] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] return self._exit_event.wait() [ 776.696759] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 776.696759] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] current.throw(*self._exc) [ 776.697428] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 776.697428] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] result = function(*args, **kwargs) [ 776.697428] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 776.697428] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] return func(*args, **kwargs) [ 776.697428] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 776.697428] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] raise e [ 776.697428] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.697428] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] nwinfo = self.network_api.allocate_for_instance( [ 776.697428] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 776.697428] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] created_port_ids = self._update_ports_for_instance( [ 776.697428] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 776.697428] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] with excutils.save_and_reraise_exception(): [ 776.697428] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.697831] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] self.force_reraise() [ 776.697831] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.697831] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] raise self.value [ 776.697831] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 776.697831] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] updated_port = self._update_port( [ 776.697831] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.697831] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] _ensure_no_port_binding_failure(port) [ 776.697831] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.697831] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] raise exception.PortBindingFailed(port_id=port['id']) [ 776.697831] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] nova.exception.PortBindingFailed: Binding failed for port 5b0c073f-e9de-4f25-ab23-376446d9aad6, please check neutron logs for more information. [ 776.697831] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] [ 776.697831] env[61970]: INFO nova.compute.manager [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Terminating instance [ 776.698192] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Acquiring lock "refresh_cache-95d428e6-fd95-4628-9053-76d218f39258" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.698192] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Acquired lock "refresh_cache-95d428e6-fd95-4628-9053-76d218f39258" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.698292] env[61970]: DEBUG nova.network.neutron [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 776.844606] env[61970]: DEBUG nova.compute.manager [req-cf4a7c93-9cbf-4dd6-9b4d-b04aac28a170 req-d190554d-1ade-41cd-8622-f7f6ea5b4685 service nova] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Received event network-changed-5b0c073f-e9de-4f25-ab23-376446d9aad6 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.844782] env[61970]: DEBUG nova.compute.manager [req-cf4a7c93-9cbf-4dd6-9b4d-b04aac28a170 req-d190554d-1ade-41cd-8622-f7f6ea5b4685 service nova] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Refreshing instance network info cache due to event network-changed-5b0c073f-e9de-4f25-ab23-376446d9aad6. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 776.844966] env[61970]: DEBUG oslo_concurrency.lockutils [req-cf4a7c93-9cbf-4dd6-9b4d-b04aac28a170 req-d190554d-1ade-41cd-8622-f7f6ea5b4685 service nova] Acquiring lock "refresh_cache-95d428e6-fd95-4628-9053-76d218f39258" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.934943] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Acquiring lock "035fb797-21c1-4af9-8e66-deee8e9d083a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.935178] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Lock "035fb797-21c1-4af9-8e66-deee8e9d083a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.044761] env[61970]: DEBUG oslo_concurrency.lockutils [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.104248] env[61970]: DEBUG nova.network.neutron [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.180322] env[61970]: DEBUG nova.network.neutron [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.219547] env[61970]: DEBUG nova.network.neutron [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.291691] env[61970]: DEBUG nova.network.neutron [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.685190] env[61970]: DEBUG oslo_concurrency.lockutils [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Releasing lock "refresh_cache-fd4274ed-bf9a-45ab-a350-a8ebbc5ad404" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.685787] env[61970]: DEBUG nova.compute.manager [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 777.685787] env[61970]: DEBUG nova.compute.manager [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 777.685787] env[61970]: DEBUG nova.network.neutron [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 777.725086] env[61970]: DEBUG nova.network.neutron [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.794040] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Releasing lock "refresh_cache-95d428e6-fd95-4628-9053-76d218f39258" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.794481] env[61970]: DEBUG nova.compute.manager [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 777.794666] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 777.795183] env[61970]: DEBUG oslo_concurrency.lockutils [req-cf4a7c93-9cbf-4dd6-9b4d-b04aac28a170 req-d190554d-1ade-41cd-8622-f7f6ea5b4685 service nova] Acquired lock "refresh_cache-95d428e6-fd95-4628-9053-76d218f39258" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.795368] env[61970]: DEBUG nova.network.neutron [req-cf4a7c93-9cbf-4dd6-9b4d-b04aac28a170 req-d190554d-1ade-41cd-8622-f7f6ea5b4685 service nova] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Refreshing network info cache for port 5b0c073f-e9de-4f25-ab23-376446d9aad6 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 777.796542] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c5e58d01-2006-478e-bec3-7d7925517355 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.809037] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82ddacaa-ae09-4f99-84d4-64e7efe36475 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.837413] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 95d428e6-fd95-4628-9053-76d218f39258 could not be found. [ 777.837640] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 777.837844] env[61970]: INFO nova.compute.manager [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Took 0.04 seconds to destroy the instance on the hypervisor. [ 777.838266] env[61970]: DEBUG oslo.service.loopingcall [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 777.841461] env[61970]: DEBUG nova.compute.manager [-] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 777.841572] env[61970]: DEBUG nova.network.neutron [-] [instance: 95d428e6-fd95-4628-9053-76d218f39258] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 777.859213] env[61970]: DEBUG nova.network.neutron [-] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.969685] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e4eff3-0424-4e1b-a166-b2db4937e9fd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.981252] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e031550e-3531-486a-a4f1-9cc73b3ca546 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.007837] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f589ed-99bf-4938-9023-e2d30e7b8f5d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.015274] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc93d8de-3cc2-488a-be6f-dcad45a1651a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.028269] env[61970]: DEBUG nova.compute.provider_tree [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.228841] env[61970]: DEBUG nova.network.neutron [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.321443] env[61970]: DEBUG nova.network.neutron [req-cf4a7c93-9cbf-4dd6-9b4d-b04aac28a170 req-d190554d-1ade-41cd-8622-f7f6ea5b4685 service nova] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.362911] env[61970]: DEBUG nova.network.neutron [-] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.411908] env[61970]: DEBUG nova.network.neutron [req-cf4a7c93-9cbf-4dd6-9b4d-b04aac28a170 req-d190554d-1ade-41cd-8622-f7f6ea5b4685 service nova] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.532564] env[61970]: DEBUG nova.scheduler.client.report [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 778.733928] env[61970]: INFO nova.compute.manager [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: fd4274ed-bf9a-45ab-a350-a8ebbc5ad404] Took 1.05 seconds to deallocate network for instance. [ 778.865395] env[61970]: INFO nova.compute.manager [-] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Took 1.02 seconds to deallocate network for instance. [ 778.867875] env[61970]: DEBUG nova.compute.claims [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 778.868082] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.917575] env[61970]: DEBUG oslo_concurrency.lockutils [req-cf4a7c93-9cbf-4dd6-9b4d-b04aac28a170 req-d190554d-1ade-41cd-8622-f7f6ea5b4685 service nova] Releasing lock "refresh_cache-95d428e6-fd95-4628-9053-76d218f39258" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.917842] env[61970]: DEBUG nova.compute.manager [req-cf4a7c93-9cbf-4dd6-9b4d-b04aac28a170 req-d190554d-1ade-41cd-8622-f7f6ea5b4685 service nova] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Received event network-vif-deleted-5b0c073f-e9de-4f25-ab23-376446d9aad6 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 779.041387] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.470s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.041957] env[61970]: DEBUG nova.compute.manager [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 779.044752] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.544s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.551464] env[61970]: DEBUG nova.compute.utils [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 779.552413] env[61970]: DEBUG nova.compute.manager [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 779.552594] env[61970]: DEBUG nova.network.neutron [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 779.630278] env[61970]: DEBUG nova.policy [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fbd50319b75e40279be1e2082a7ecb4c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '740d49b85c2f43d3a4a19d20df59da70', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 779.771878] env[61970]: INFO nova.scheduler.client.report [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Deleted allocations for instance fd4274ed-bf9a-45ab-a350-a8ebbc5ad404 [ 780.056571] env[61970]: DEBUG nova.compute.manager [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 780.071325] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance e74c9188-4f8a-4071-bc4f-c2be91fec3f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 780.071580] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 01ee68e2-4310-477e-8adb-6ad67d6a65e7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 780.071715] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 95d428e6-fd95-4628-9053-76d218f39258 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 780.071831] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 3f0767a0-9925-4cb9-8032-b5ca071cd572 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 780.127311] env[61970]: DEBUG nova.network.neutron [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Successfully created port: f8eba464-80eb-4d70-af1e-f0a5fc33895b {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.279958] env[61970]: DEBUG oslo_concurrency.lockutils [None req-47f1eba8-15f7-49ab-9c5c-6c8b7535cbe3 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "fd4274ed-bf9a-45ab-a350-a8ebbc5ad404" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.469s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.577186] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 780.786078] env[61970]: DEBUG nova.compute.manager [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 781.074287] env[61970]: DEBUG nova.compute.manager [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 781.081876] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 3528d9b6-f3fb-43e9-be27-113d887a3414 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 781.095815] env[61970]: DEBUG nova.compute.manager [req-f6025874-d8dc-41ec-82b1-5f92619e5f6a req-c6447ecf-3a4e-46b5-81b4-7ec7b2854d6e service nova] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Received event network-changed-f8eba464-80eb-4d70-af1e-f0a5fc33895b {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 781.095943] env[61970]: DEBUG nova.compute.manager [req-f6025874-d8dc-41ec-82b1-5f92619e5f6a req-c6447ecf-3a4e-46b5-81b4-7ec7b2854d6e service nova] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Refreshing instance network info cache due to event network-changed-f8eba464-80eb-4d70-af1e-f0a5fc33895b. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 781.096185] env[61970]: DEBUG oslo_concurrency.lockutils [req-f6025874-d8dc-41ec-82b1-5f92619e5f6a req-c6447ecf-3a4e-46b5-81b4-7ec7b2854d6e service nova] Acquiring lock "refresh_cache-3f0767a0-9925-4cb9-8032-b5ca071cd572" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.096330] env[61970]: DEBUG oslo_concurrency.lockutils [req-f6025874-d8dc-41ec-82b1-5f92619e5f6a req-c6447ecf-3a4e-46b5-81b4-7ec7b2854d6e service nova] Acquired lock "refresh_cache-3f0767a0-9925-4cb9-8032-b5ca071cd572" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.096524] env[61970]: DEBUG nova.network.neutron [req-f6025874-d8dc-41ec-82b1-5f92619e5f6a req-c6447ecf-3a4e-46b5-81b4-7ec7b2854d6e service nova] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Refreshing network info cache for port f8eba464-80eb-4d70-af1e-f0a5fc33895b {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 781.102893] env[61970]: DEBUG nova.virt.hardware [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 781.103159] env[61970]: DEBUG nova.virt.hardware [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 781.103335] env[61970]: DEBUG nova.virt.hardware [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.103539] env[61970]: DEBUG nova.virt.hardware [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 781.103653] env[61970]: DEBUG nova.virt.hardware [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.103794] env[61970]: DEBUG nova.virt.hardware [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 781.103998] env[61970]: DEBUG nova.virt.hardware [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 781.104305] env[61970]: DEBUG nova.virt.hardware [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 781.104513] env[61970]: DEBUG nova.virt.hardware [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 781.104686] env[61970]: DEBUG nova.virt.hardware [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 781.104857] env[61970]: DEBUG nova.virt.hardware [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 781.106142] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7407d710-1f76-4091-b9e5-44c99211cb99 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.123102] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39702674-c25f-48d4-9a9c-280f31264b20 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.170103] env[61970]: DEBUG nova.network.neutron [req-f6025874-d8dc-41ec-82b1-5f92619e5f6a req-c6447ecf-3a4e-46b5-81b4-7ec7b2854d6e service nova] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.310426] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.319212] env[61970]: DEBUG nova.network.neutron [req-f6025874-d8dc-41ec-82b1-5f92619e5f6a req-c6447ecf-3a4e-46b5-81b4-7ec7b2854d6e service nova] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.396048] env[61970]: ERROR nova.compute.manager [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f8eba464-80eb-4d70-af1e-f0a5fc33895b, please check neutron logs for more information. [ 781.396048] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 781.396048] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.396048] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 781.396048] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.396048] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 781.396048] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.396048] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 781.396048] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.396048] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 781.396048] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.396048] env[61970]: ERROR nova.compute.manager raise self.value [ 781.396048] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.396048] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 781.396048] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.396048] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 781.396523] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.396523] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 781.396523] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f8eba464-80eb-4d70-af1e-f0a5fc33895b, please check neutron logs for more information. [ 781.396523] env[61970]: ERROR nova.compute.manager [ 781.396523] env[61970]: Traceback (most recent call last): [ 781.396523] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 781.396523] env[61970]: listener.cb(fileno) [ 781.396523] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.396523] env[61970]: result = function(*args, **kwargs) [ 781.396523] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 781.396523] env[61970]: return func(*args, **kwargs) [ 781.396523] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.396523] env[61970]: raise e [ 781.396523] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.396523] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 781.396523] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.396523] env[61970]: created_port_ids = self._update_ports_for_instance( [ 781.396523] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.396523] env[61970]: with excutils.save_and_reraise_exception(): [ 781.396523] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.396523] env[61970]: self.force_reraise() [ 781.396523] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.396523] env[61970]: raise self.value [ 781.396523] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.396523] env[61970]: updated_port = self._update_port( [ 781.396523] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.396523] env[61970]: _ensure_no_port_binding_failure(port) [ 781.396523] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.396523] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 781.397503] env[61970]: nova.exception.PortBindingFailed: Binding failed for port f8eba464-80eb-4d70-af1e-f0a5fc33895b, please check neutron logs for more information. [ 781.397503] env[61970]: Removing descriptor: 15 [ 781.397503] env[61970]: ERROR nova.compute.manager [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f8eba464-80eb-4d70-af1e-f0a5fc33895b, please check neutron logs for more information. [ 781.397503] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Traceback (most recent call last): [ 781.397503] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 781.397503] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] yield resources [ 781.397503] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 781.397503] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] self.driver.spawn(context, instance, image_meta, [ 781.397503] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 781.397503] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.397503] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.397503] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] vm_ref = self.build_virtual_machine(instance, [ 781.397897] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.397897] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.397897] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.397897] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] for vif in network_info: [ 781.397897] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 781.397897] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] return self._sync_wrapper(fn, *args, **kwargs) [ 781.397897] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 781.397897] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] self.wait() [ 781.397897] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 781.397897] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] self[:] = self._gt.wait() [ 781.397897] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.397897] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] return self._exit_event.wait() [ 781.397897] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 781.398349] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] result = hub.switch() [ 781.398349] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 781.398349] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] return self.greenlet.switch() [ 781.398349] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.398349] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] result = function(*args, **kwargs) [ 781.398349] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 781.398349] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] return func(*args, **kwargs) [ 781.398349] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.398349] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] raise e [ 781.398349] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.398349] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] nwinfo = self.network_api.allocate_for_instance( [ 781.398349] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.398349] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] created_port_ids = self._update_ports_for_instance( [ 781.398776] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.398776] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] with excutils.save_and_reraise_exception(): [ 781.398776] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.398776] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] self.force_reraise() [ 781.398776] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.398776] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] raise self.value [ 781.398776] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.398776] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] updated_port = self._update_port( [ 781.398776] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.398776] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] _ensure_no_port_binding_failure(port) [ 781.398776] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.398776] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] raise exception.PortBindingFailed(port_id=port['id']) [ 781.399243] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] nova.exception.PortBindingFailed: Binding failed for port f8eba464-80eb-4d70-af1e-f0a5fc33895b, please check neutron logs for more information. [ 781.399243] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] [ 781.399243] env[61970]: INFO nova.compute.manager [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Terminating instance [ 781.399243] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Acquiring lock "refresh_cache-3f0767a0-9925-4cb9-8032-b5ca071cd572" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.480503] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "2d280f75-ed84-4db3-9067-8ccc719b0d2e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.480826] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "2d280f75-ed84-4db3-9067-8ccc719b0d2e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.584837] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 07eeb310-312b-4f32-bd5f-7b254cff0fe1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 781.822654] env[61970]: DEBUG oslo_concurrency.lockutils [req-f6025874-d8dc-41ec-82b1-5f92619e5f6a req-c6447ecf-3a4e-46b5-81b4-7ec7b2854d6e service nova] Releasing lock "refresh_cache-3f0767a0-9925-4cb9-8032-b5ca071cd572" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.823063] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Acquired lock "refresh_cache-3f0767a0-9925-4cb9-8032-b5ca071cd572" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.823358] env[61970]: DEBUG nova.network.neutron [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 782.087682] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance a8ecafbe-f1e1-49f1-945d-ef8b6d15b709 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 782.347555] env[61970]: DEBUG nova.network.neutron [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.440848] env[61970]: DEBUG nova.network.neutron [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.590856] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 782.944611] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Releasing lock "refresh_cache-3f0767a0-9925-4cb9-8032-b5ca071cd572" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.945048] env[61970]: DEBUG nova.compute.manager [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 782.945243] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 782.945619] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f68981ed-8797-4c14-8041-3a72a7b233e0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.954893] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-724ef11c-9ed1-4109-b21c-e5946d95f384 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.977853] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3f0767a0-9925-4cb9-8032-b5ca071cd572 could not be found. [ 782.977853] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 782.977853] env[61970]: INFO nova.compute.manager [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Took 0.03 seconds to destroy the instance on the hypervisor. [ 782.977853] env[61970]: DEBUG oslo.service.loopingcall [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.978125] env[61970]: DEBUG nova.compute.manager [-] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.978125] env[61970]: DEBUG nova.network.neutron [-] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 782.993114] env[61970]: DEBUG nova.network.neutron [-] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.094760] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance a78b63a5-6bb8-4271-90d8-1e86fb29db4f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 783.126026] env[61970]: DEBUG nova.compute.manager [req-7a0ab17d-2176-47f4-a9c6-0b7dfb39d08d req-4005a329-4a55-4504-b216-d663f4535036 service nova] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Received event network-vif-deleted-f8eba464-80eb-4d70-af1e-f0a5fc33895b {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 783.495771] env[61970]: DEBUG nova.network.neutron [-] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.599561] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 27996960-d2d4-496c-b52d-5d7312aed040 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 784.001447] env[61970]: INFO nova.compute.manager [-] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Took 1.02 seconds to deallocate network for instance. [ 784.005443] env[61970]: DEBUG nova.compute.claims [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 784.005443] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.104644] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance c5c253b6-ebab-4fea-bcc4-f3bb3960070b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 784.609203] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 5d4d97e1-5812-4884-a1ea-6eb0cee591c0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 785.113360] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 6345d239-fbae-4f3a-9f61-79e10539ff0b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 785.617156] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance e2f51299-56fe-46cc-9e08-2246178b9db7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 786.120949] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance b19cb516-b163-4bed-ba5b-139a0a18fc05 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 786.623865] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance b1cc1cc2-15d6-459d-9529-e592ddb225ac has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 787.128812] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance b385407b-1bdd-4c53-907c-cb4c8ce16cc7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 787.633234] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 788.140205] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance d09dbde2-5daa-4ae9-9fd5-bfbb03fba136 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 788.482810] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.483104] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.643244] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance e1415314-f137-4ee3-a065-3d875ef9e7ff has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 789.148258] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 035fb797-21c1-4af9-8e66-deee8e9d083a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 789.148698] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 789.148895] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 789.410604] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bbff2b6-acd3-4b17-a4e9-93fa986fa9de {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.418248] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0668497a-0f16-4a45-b134-8bb664417ac8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.446456] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf4f9ab-485e-4ed7-901c-829d382a1f98 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.453225] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd00840-c682-4800-b640-cd563e0cc487 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.465744] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.969133] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 790.474691] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 790.474955] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 11.430s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.475260] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.609s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.478502] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 790.478641] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Cleaning up deleted instances {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 790.985215] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] There are 4 instances to clean {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 790.985491] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 915c3b6c-640b-430f-b264-40bf85c642b9] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 791.070954] env[61970]: DEBUG nova.scheduler.client.report [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Refreshing inventories for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 791.087598] env[61970]: DEBUG nova.scheduler.client.report [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Updating ProviderTree inventory for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 791.087824] env[61970]: DEBUG nova.compute.provider_tree [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 791.099142] env[61970]: DEBUG nova.scheduler.client.report [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Refreshing aggregate associations for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa, aggregates: None {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 791.117235] env[61970]: DEBUG nova.scheduler.client.report [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Refreshing trait associations for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 791.367218] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf43f148-4bb3-4f1f-bf58-46207075b884 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.374656] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf62a90-3ea4-4f01-9b62-83e17564d477 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.403286] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73db96d2-cb80-458e-af20-e109f1f4d34c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.410083] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bec7dd4-26c6-4cee-9cb2-86fb6574035f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.422560] env[61970]: DEBUG nova.compute.provider_tree [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.489270] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 9e166ffe-1ada-43d2-9ac1-56ad0cf7cae8] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 791.925447] env[61970]: DEBUG nova.scheduler.client.report [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 791.991965] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 8881014c-8505-4e75-8c31-e25b6a8d2cd4] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 792.429909] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.954s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.430566] env[61970]: ERROR nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 774d7fd3-a724-4311-a158-c727527f9cb3, please check neutron logs for more information. [ 792.430566] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Traceback (most recent call last): [ 792.430566] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 792.430566] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] self.driver.spawn(context, instance, image_meta, [ 792.430566] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 792.430566] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 792.430566] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 792.430566] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] vm_ref = self.build_virtual_machine(instance, [ 792.430566] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 792.430566] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] vif_infos = vmwarevif.get_vif_info(self._session, [ 792.430566] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 792.430831] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] for vif in network_info: [ 792.430831] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 792.430831] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] return self._sync_wrapper(fn, *args, **kwargs) [ 792.430831] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 792.430831] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] self.wait() [ 792.430831] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 792.430831] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] self[:] = self._gt.wait() [ 792.430831] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 792.430831] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] return self._exit_event.wait() [ 792.430831] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 792.430831] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] current.throw(*self._exc) [ 792.430831] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 792.430831] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] result = function(*args, **kwargs) [ 792.431111] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 792.431111] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] return func(*args, **kwargs) [ 792.431111] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 792.431111] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] raise e [ 792.431111] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 792.431111] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] nwinfo = self.network_api.allocate_for_instance( [ 792.431111] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 792.431111] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] created_port_ids = self._update_ports_for_instance( [ 792.431111] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 792.431111] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] with excutils.save_and_reraise_exception(): [ 792.431111] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 792.431111] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] self.force_reraise() [ 792.431111] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 792.431413] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] raise self.value [ 792.431413] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 792.431413] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] updated_port = self._update_port( [ 792.431413] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 792.431413] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] _ensure_no_port_binding_failure(port) [ 792.431413] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 792.431413] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] raise exception.PortBindingFailed(port_id=port['id']) [ 792.431413] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] nova.exception.PortBindingFailed: Binding failed for port 774d7fd3-a724-4311-a158-c727527f9cb3, please check neutron logs for more information. [ 792.431413] env[61970]: ERROR nova.compute.manager [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] [ 792.431413] env[61970]: DEBUG nova.compute.utils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Binding failed for port 774d7fd3-a724-4311-a158-c727527f9cb3, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 792.433047] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.219s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.434514] env[61970]: INFO nova.compute.claims [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 792.437386] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Build of instance e74c9188-4f8a-4071-bc4f-c2be91fec3f7 was re-scheduled: Binding failed for port 774d7fd3-a724-4311-a158-c727527f9cb3, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 792.437627] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 792.437931] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "refresh_cache-e74c9188-4f8a-4071-bc4f-c2be91fec3f7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.438075] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquired lock "refresh_cache-e74c9188-4f8a-4071-bc4f-c2be91fec3f7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.438240] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 792.495221] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: c4d3b145-84be-4d19-afc7-4c29fa193760] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 792.958576] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.999912] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 793.000470] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Cleaning up deleted instances with incomplete migration {{(pid=61970) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 793.052445] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.502378] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 793.556049] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Releasing lock "refresh_cache-e74c9188-4f8a-4071-bc4f-c2be91fec3f7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.556049] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 793.556049] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 793.556049] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 793.571558] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.700995] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56bc7d69-cf5e-4d24-8a47-866447e0b12d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.708755] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86c567eb-458d-4a68-a72a-95f010db903c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.737609] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b9bcb6-ff4b-4df9-9184-cc160b6bc570 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.744617] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a55870-e905-4b25-a5a5-8117b90f8443 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.760233] env[61970]: DEBUG nova.compute.provider_tree [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.076032] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.261402] env[61970]: DEBUG nova.scheduler.client.report [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 794.578219] env[61970]: INFO nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e74c9188-4f8a-4071-bc4f-c2be91fec3f7] Took 1.02 seconds to deallocate network for instance. [ 794.765904] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.766450] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 794.769253] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.713s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.771292] env[61970]: INFO nova.compute.claims [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 795.275393] env[61970]: DEBUG nova.compute.utils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 795.278857] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 795.278857] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 795.325762] env[61970]: DEBUG nova.policy [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc95d6e06df24083afd6947e861f6b36', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'da5b3815af8d455b888c541d8b0d7b32', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 795.589169] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Successfully created port: 3a2c383f-766e-454a-bddf-a884d879da19 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 795.611448] env[61970]: INFO nova.scheduler.client.report [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Deleted allocations for instance e74c9188-4f8a-4071-bc4f-c2be91fec3f7 [ 795.780063] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 796.051158] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0504518-5dad-4b87-8692-d5848dc81927 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.059897] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b9c177-0602-43ec-87ec-4f983ed38f0a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.090315] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b8ad996-51a2-45a5-add9-5bec4f047ed1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.097117] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb5f44d-af22-4ad7-8491-764555a9bc70 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.109778] env[61970]: DEBUG nova.compute.provider_tree [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.119101] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "e74c9188-4f8a-4071-bc4f-c2be91fec3f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.708s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.304150] env[61970]: DEBUG nova.compute.manager [req-4e9d2ac8-1213-46f4-ad83-6ce25637c493 req-a1958d7e-1ac1-4af6-8be2-ba77f7a4dcbd service nova] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Received event network-changed-3a2c383f-766e-454a-bddf-a884d879da19 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 796.304436] env[61970]: DEBUG nova.compute.manager [req-4e9d2ac8-1213-46f4-ad83-6ce25637c493 req-a1958d7e-1ac1-4af6-8be2-ba77f7a4dcbd service nova] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Refreshing instance network info cache due to event network-changed-3a2c383f-766e-454a-bddf-a884d879da19. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 796.305094] env[61970]: DEBUG oslo_concurrency.lockutils [req-4e9d2ac8-1213-46f4-ad83-6ce25637c493 req-a1958d7e-1ac1-4af6-8be2-ba77f7a4dcbd service nova] Acquiring lock "refresh_cache-5ce0137d-a22e-4f22-9d3f-ed4812e67ec7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.305094] env[61970]: DEBUG oslo_concurrency.lockutils [req-4e9d2ac8-1213-46f4-ad83-6ce25637c493 req-a1958d7e-1ac1-4af6-8be2-ba77f7a4dcbd service nova] Acquired lock "refresh_cache-5ce0137d-a22e-4f22-9d3f-ed4812e67ec7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.305094] env[61970]: DEBUG nova.network.neutron [req-4e9d2ac8-1213-46f4-ad83-6ce25637c493 req-a1958d7e-1ac1-4af6-8be2-ba77f7a4dcbd service nova] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Refreshing network info cache for port 3a2c383f-766e-454a-bddf-a884d879da19 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 796.504996] env[61970]: ERROR nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3a2c383f-766e-454a-bddf-a884d879da19, please check neutron logs for more information. [ 796.504996] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 796.504996] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 796.504996] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 796.504996] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 796.504996] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 796.504996] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 796.504996] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 796.504996] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 796.504996] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 796.504996] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 796.504996] env[61970]: ERROR nova.compute.manager raise self.value [ 796.504996] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 796.504996] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 796.504996] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 796.504996] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 796.505463] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 796.505463] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 796.505463] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3a2c383f-766e-454a-bddf-a884d879da19, please check neutron logs for more information. [ 796.505463] env[61970]: ERROR nova.compute.manager [ 796.505463] env[61970]: Traceback (most recent call last): [ 796.505463] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 796.505463] env[61970]: listener.cb(fileno) [ 796.505463] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 796.505463] env[61970]: result = function(*args, **kwargs) [ 796.505463] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 796.505463] env[61970]: return func(*args, **kwargs) [ 796.505463] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 796.505463] env[61970]: raise e [ 796.505463] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 796.505463] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 796.505463] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 796.505463] env[61970]: created_port_ids = self._update_ports_for_instance( [ 796.505463] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 796.505463] env[61970]: with excutils.save_and_reraise_exception(): [ 796.505463] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 796.505463] env[61970]: self.force_reraise() [ 796.505463] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 796.505463] env[61970]: raise self.value [ 796.505463] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 796.505463] env[61970]: updated_port = self._update_port( [ 796.505463] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 796.505463] env[61970]: _ensure_no_port_binding_failure(port) [ 796.505463] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 796.505463] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 796.506320] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 3a2c383f-766e-454a-bddf-a884d879da19, please check neutron logs for more information. [ 796.506320] env[61970]: Removing descriptor: 15 [ 796.613029] env[61970]: DEBUG nova.scheduler.client.report [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 796.621338] env[61970]: DEBUG nova.compute.manager [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 796.792482] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 796.817052] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 796.817304] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 796.817458] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 796.817635] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 796.817777] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 796.817921] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 796.818142] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 796.818299] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 796.818462] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 796.818621] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 796.818788] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 796.819646] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afc56b96-fcba-4325-b382-4908cb14f8eb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.823424] env[61970]: DEBUG nova.network.neutron [req-4e9d2ac8-1213-46f4-ad83-6ce25637c493 req-a1958d7e-1ac1-4af6-8be2-ba77f7a4dcbd service nova] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.828227] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9cde88-87aa-40b2-b997-9529a9023b92 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.842821] env[61970]: ERROR nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3a2c383f-766e-454a-bddf-a884d879da19, please check neutron logs for more information. [ 796.842821] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Traceback (most recent call last): [ 796.842821] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 796.842821] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] yield resources [ 796.842821] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 796.842821] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] self.driver.spawn(context, instance, image_meta, [ 796.842821] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 796.842821] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 796.842821] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 796.842821] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] vm_ref = self.build_virtual_machine(instance, [ 796.842821] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 796.843181] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] vif_infos = vmwarevif.get_vif_info(self._session, [ 796.843181] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 796.843181] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] for vif in network_info: [ 796.843181] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 796.843181] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] return self._sync_wrapper(fn, *args, **kwargs) [ 796.843181] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 796.843181] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] self.wait() [ 796.843181] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 796.843181] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] self[:] = self._gt.wait() [ 796.843181] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 796.843181] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] return self._exit_event.wait() [ 796.843181] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 796.843181] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] current.throw(*self._exc) [ 796.843493] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 796.843493] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] result = function(*args, **kwargs) [ 796.843493] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 796.843493] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] return func(*args, **kwargs) [ 796.843493] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 796.843493] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] raise e [ 796.843493] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 796.843493] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] nwinfo = self.network_api.allocate_for_instance( [ 796.843493] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 796.843493] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] created_port_ids = self._update_ports_for_instance( [ 796.843493] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 796.843493] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] with excutils.save_and_reraise_exception(): [ 796.843493] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 796.843761] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] self.force_reraise() [ 796.843761] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 796.843761] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] raise self.value [ 796.843761] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 796.843761] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] updated_port = self._update_port( [ 796.843761] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 796.843761] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] _ensure_no_port_binding_failure(port) [ 796.843761] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 796.843761] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] raise exception.PortBindingFailed(port_id=port['id']) [ 796.843761] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] nova.exception.PortBindingFailed: Binding failed for port 3a2c383f-766e-454a-bddf-a884d879da19, please check neutron logs for more information. [ 796.843761] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] [ 796.843761] env[61970]: INFO nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Terminating instance [ 796.845685] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquiring lock "refresh_cache-5ce0137d-a22e-4f22-9d3f-ed4812e67ec7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.893753] env[61970]: DEBUG nova.network.neutron [req-4e9d2ac8-1213-46f4-ad83-6ce25637c493 req-a1958d7e-1ac1-4af6-8be2-ba77f7a4dcbd service nova] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.117797] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.348s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.118372] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 797.121031] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.255s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.122822] env[61970]: INFO nova.compute.claims [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 797.142951] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.396160] env[61970]: DEBUG oslo_concurrency.lockutils [req-4e9d2ac8-1213-46f4-ad83-6ce25637c493 req-a1958d7e-1ac1-4af6-8be2-ba77f7a4dcbd service nova] Releasing lock "refresh_cache-5ce0137d-a22e-4f22-9d3f-ed4812e67ec7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.396566] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquired lock "refresh_cache-5ce0137d-a22e-4f22-9d3f-ed4812e67ec7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.396751] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 797.626876] env[61970]: DEBUG nova.compute.utils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 797.628321] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 797.628473] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 797.680441] env[61970]: DEBUG nova.policy [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc95d6e06df24083afd6947e861f6b36', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'da5b3815af8d455b888c541d8b0d7b32', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 797.925877] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 797.953481] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Successfully created port: c2e6332a-9fc1-4dbb-aea3-77565735c463 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 798.057131] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.134066] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 798.334280] env[61970]: DEBUG nova.compute.manager [req-41b4e42c-a004-4f85-bc2a-e053c94bb552 req-8ff89b8e-881e-4c0d-bac0-70aca01a660f service nova] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Received event network-vif-deleted-3a2c383f-766e-454a-bddf-a884d879da19 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 798.497998] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2d6ce6f-7fe5-4fb9-bff3-17d50670b2ee {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.507023] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ad15ea5-befa-42e3-865a-9cf146ca02e0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.542861] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc383bd2-7077-4f5f-99ba-03c42af5d668 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.550520] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03261fae-9365-4ede-ab01-526bffaa4d00 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.566981] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Releasing lock "refresh_cache-5ce0137d-a22e-4f22-9d3f-ed4812e67ec7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.567427] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 798.567621] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 798.568091] env[61970]: DEBUG nova.compute.provider_tree [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.569425] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8b3a1a86-4aab-4651-ac04-777c7da48ed2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.577899] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df47143a-a07c-43b8-a01e-77760d27d09e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.599650] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7 could not be found. [ 798.599872] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 798.600098] env[61970]: INFO nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Took 0.03 seconds to destroy the instance on the hypervisor. [ 798.600425] env[61970]: DEBUG oslo.service.loopingcall [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.600557] env[61970]: DEBUG nova.compute.manager [-] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 798.600648] env[61970]: DEBUG nova.network.neutron [-] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 798.615726] env[61970]: DEBUG nova.network.neutron [-] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 798.885324] env[61970]: ERROR nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c2e6332a-9fc1-4dbb-aea3-77565735c463, please check neutron logs for more information. [ 798.885324] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 798.885324] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 798.885324] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 798.885324] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 798.885324] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 798.885324] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 798.885324] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 798.885324] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 798.885324] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 798.885324] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 798.885324] env[61970]: ERROR nova.compute.manager raise self.value [ 798.885324] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 798.885324] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 798.885324] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 798.885324] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 798.886131] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 798.886131] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 798.886131] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c2e6332a-9fc1-4dbb-aea3-77565735c463, please check neutron logs for more information. [ 798.886131] env[61970]: ERROR nova.compute.manager [ 798.886131] env[61970]: Traceback (most recent call last): [ 798.886131] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 798.886131] env[61970]: listener.cb(fileno) [ 798.886131] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 798.886131] env[61970]: result = function(*args, **kwargs) [ 798.886131] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 798.886131] env[61970]: return func(*args, **kwargs) [ 798.886131] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 798.886131] env[61970]: raise e [ 798.886131] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 798.886131] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 798.886131] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 798.886131] env[61970]: created_port_ids = self._update_ports_for_instance( [ 798.886131] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 798.886131] env[61970]: with excutils.save_and_reraise_exception(): [ 798.886131] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 798.886131] env[61970]: self.force_reraise() [ 798.886131] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 798.886131] env[61970]: raise self.value [ 798.886131] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 798.886131] env[61970]: updated_port = self._update_port( [ 798.886131] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 798.886131] env[61970]: _ensure_no_port_binding_failure(port) [ 798.886131] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 798.886131] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 798.886844] env[61970]: nova.exception.PortBindingFailed: Binding failed for port c2e6332a-9fc1-4dbb-aea3-77565735c463, please check neutron logs for more information. [ 798.886844] env[61970]: Removing descriptor: 15 [ 799.072434] env[61970]: DEBUG nova.scheduler.client.report [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 799.120646] env[61970]: DEBUG nova.network.neutron [-] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.149646] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 799.175330] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 799.175581] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 799.175736] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 799.175913] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 799.176070] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 799.176223] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 799.176425] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 799.176582] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 799.176751] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 799.176912] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 799.177098] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 799.177990] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb71e6d4-3eae-473d-b419-e68cc7b4acb6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.186146] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0182465-9915-4bdd-b804-f0f0d2d33998 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.200242] env[61970]: ERROR nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c2e6332a-9fc1-4dbb-aea3-77565735c463, please check neutron logs for more information. [ 799.200242] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Traceback (most recent call last): [ 799.200242] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 799.200242] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] yield resources [ 799.200242] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 799.200242] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] self.driver.spawn(context, instance, image_meta, [ 799.200242] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 799.200242] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] self._vmops.spawn(context, instance, image_meta, injected_files, [ 799.200242] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 799.200242] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] vm_ref = self.build_virtual_machine(instance, [ 799.200242] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 799.200508] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] vif_infos = vmwarevif.get_vif_info(self._session, [ 799.200508] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 799.200508] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] for vif in network_info: [ 799.200508] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 799.200508] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] return self._sync_wrapper(fn, *args, **kwargs) [ 799.200508] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 799.200508] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] self.wait() [ 799.200508] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 799.200508] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] self[:] = self._gt.wait() [ 799.200508] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 799.200508] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] return self._exit_event.wait() [ 799.200508] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 799.200508] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] current.throw(*self._exc) [ 799.200770] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 799.200770] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] result = function(*args, **kwargs) [ 799.200770] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 799.200770] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] return func(*args, **kwargs) [ 799.200770] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 799.200770] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] raise e [ 799.200770] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 799.200770] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] nwinfo = self.network_api.allocate_for_instance( [ 799.200770] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 799.200770] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] created_port_ids = self._update_ports_for_instance( [ 799.200770] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 799.200770] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] with excutils.save_and_reraise_exception(): [ 799.200770] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 799.201100] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] self.force_reraise() [ 799.201100] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 799.201100] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] raise self.value [ 799.201100] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 799.201100] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] updated_port = self._update_port( [ 799.201100] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 799.201100] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] _ensure_no_port_binding_failure(port) [ 799.201100] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 799.201100] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] raise exception.PortBindingFailed(port_id=port['id']) [ 799.201100] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] nova.exception.PortBindingFailed: Binding failed for port c2e6332a-9fc1-4dbb-aea3-77565735c463, please check neutron logs for more information. [ 799.201100] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] [ 799.201100] env[61970]: INFO nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Terminating instance [ 799.202559] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquiring lock "refresh_cache-3528d9b6-f3fb-43e9-be27-113d887a3414" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.202724] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquired lock "refresh_cache-3528d9b6-f3fb-43e9-be27-113d887a3414" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.202930] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 799.576983] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.577559] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 799.580320] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.890s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.581836] env[61970]: INFO nova.compute.claims [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 799.622559] env[61970]: INFO nova.compute.manager [-] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Took 1.02 seconds to deallocate network for instance. [ 799.624825] env[61970]: DEBUG nova.compute.claims [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 799.625013] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.723625] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 799.807709] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.086488] env[61970]: DEBUG nova.compute.utils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 800.090432] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 800.090432] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 800.134829] env[61970]: DEBUG nova.policy [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc95d6e06df24083afd6947e861f6b36', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'da5b3815af8d455b888c541d8b0d7b32', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 800.310303] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Releasing lock "refresh_cache-3528d9b6-f3fb-43e9-be27-113d887a3414" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.310789] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 800.310976] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 800.311288] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ad433aae-1271-441f-9ee0-8453d9388ce6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.320289] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5477e537-e399-48bc-965a-c34e251f9470 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.342410] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3528d9b6-f3fb-43e9-be27-113d887a3414 could not be found. [ 800.342410] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 800.342410] env[61970]: INFO nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Took 0.03 seconds to destroy the instance on the hypervisor. [ 800.342410] env[61970]: DEBUG oslo.service.loopingcall [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.342410] env[61970]: DEBUG nova.compute.manager [-] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 800.342410] env[61970]: DEBUG nova.network.neutron [-] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 800.357509] env[61970]: DEBUG nova.compute.manager [req-18be19f5-75d0-49db-a675-4ab0ed7c4a87 req-611fd97b-43e9-46c3-9d73-28b3f2d3bcb2 service nova] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Received event network-changed-c2e6332a-9fc1-4dbb-aea3-77565735c463 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 800.357690] env[61970]: DEBUG nova.compute.manager [req-18be19f5-75d0-49db-a675-4ab0ed7c4a87 req-611fd97b-43e9-46c3-9d73-28b3f2d3bcb2 service nova] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Refreshing instance network info cache due to event network-changed-c2e6332a-9fc1-4dbb-aea3-77565735c463. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 800.357896] env[61970]: DEBUG oslo_concurrency.lockutils [req-18be19f5-75d0-49db-a675-4ab0ed7c4a87 req-611fd97b-43e9-46c3-9d73-28b3f2d3bcb2 service nova] Acquiring lock "refresh_cache-3528d9b6-f3fb-43e9-be27-113d887a3414" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.358050] env[61970]: DEBUG oslo_concurrency.lockutils [req-18be19f5-75d0-49db-a675-4ab0ed7c4a87 req-611fd97b-43e9-46c3-9d73-28b3f2d3bcb2 service nova] Acquired lock "refresh_cache-3528d9b6-f3fb-43e9-be27-113d887a3414" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.358215] env[61970]: DEBUG nova.network.neutron [req-18be19f5-75d0-49db-a675-4ab0ed7c4a87 req-611fd97b-43e9-46c3-9d73-28b3f2d3bcb2 service nova] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Refreshing network info cache for port c2e6332a-9fc1-4dbb-aea3-77565735c463 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 800.368500] env[61970]: DEBUG nova.network.neutron [-] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.412764] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Successfully created port: 7434460f-bcc3-4312-8a89-15154d936f92 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 800.591180] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 800.857622] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841b9d4c-9604-4b34-900a-7d1693f9eb8f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.867175] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecad0d2d-42f8-493a-ab5e-534403aabf13 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.870291] env[61970]: DEBUG nova.network.neutron [-] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.898275] env[61970]: DEBUG nova.network.neutron [req-18be19f5-75d0-49db-a675-4ab0ed7c4a87 req-611fd97b-43e9-46c3-9d73-28b3f2d3bcb2 service nova] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.900558] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb52b082-225b-4199-84ea-640f43fb5a47 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.909355] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a9bc28-2309-48bf-aec4-e3d969ea8bb8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.629649] env[61970]: INFO nova.compute.manager [-] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Took 1.29 seconds to deallocate network for instance. [ 801.630412] env[61970]: DEBUG nova.compute.provider_tree [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.636990] env[61970]: DEBUG nova.compute.claims [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 801.637168] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.716378] env[61970]: DEBUG nova.network.neutron [req-18be19f5-75d0-49db-a675-4ab0ed7c4a87 req-611fd97b-43e9-46c3-9d73-28b3f2d3bcb2 service nova] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.891570] env[61970]: ERROR nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7434460f-bcc3-4312-8a89-15154d936f92, please check neutron logs for more information. [ 801.891570] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 801.891570] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 801.891570] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 801.891570] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 801.891570] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 801.891570] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 801.891570] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 801.891570] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 801.891570] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 801.891570] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 801.891570] env[61970]: ERROR nova.compute.manager raise self.value [ 801.891570] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 801.891570] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 801.891570] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 801.891570] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 801.892070] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 801.892070] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 801.892070] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7434460f-bcc3-4312-8a89-15154d936f92, please check neutron logs for more information. [ 801.892070] env[61970]: ERROR nova.compute.manager [ 801.892070] env[61970]: Traceback (most recent call last): [ 801.892070] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 801.892070] env[61970]: listener.cb(fileno) [ 801.892070] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 801.892070] env[61970]: result = function(*args, **kwargs) [ 801.892070] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 801.892070] env[61970]: return func(*args, **kwargs) [ 801.892070] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 801.892070] env[61970]: raise e [ 801.892070] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 801.892070] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 801.892070] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 801.892070] env[61970]: created_port_ids = self._update_ports_for_instance( [ 801.892070] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 801.892070] env[61970]: with excutils.save_and_reraise_exception(): [ 801.892070] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 801.892070] env[61970]: self.force_reraise() [ 801.892070] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 801.892070] env[61970]: raise self.value [ 801.892070] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 801.892070] env[61970]: updated_port = self._update_port( [ 801.892070] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 801.892070] env[61970]: _ensure_no_port_binding_failure(port) [ 801.892070] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 801.892070] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 801.892857] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 7434460f-bcc3-4312-8a89-15154d936f92, please check neutron logs for more information. [ 801.892857] env[61970]: Removing descriptor: 15 [ 802.136896] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 802.140099] env[61970]: DEBUG nova.scheduler.client.report [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 802.162497] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.162751] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.163043] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.163196] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.163343] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.163489] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.163719] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.163897] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.164077] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.164242] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.164412] env[61970]: DEBUG nova.virt.hardware [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.165512] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca172e07-9992-439e-a2ee-0534c8803fea {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.173561] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b9827e-325f-4e75-bc1e-2a9a7f774ae1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.187188] env[61970]: ERROR nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7434460f-bcc3-4312-8a89-15154d936f92, please check neutron logs for more information. [ 802.187188] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Traceback (most recent call last): [ 802.187188] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 802.187188] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] yield resources [ 802.187188] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 802.187188] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] self.driver.spawn(context, instance, image_meta, [ 802.187188] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 802.187188] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 802.187188] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 802.187188] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] vm_ref = self.build_virtual_machine(instance, [ 802.187188] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 802.187589] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] vif_infos = vmwarevif.get_vif_info(self._session, [ 802.187589] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 802.187589] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] for vif in network_info: [ 802.187589] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 802.187589] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] return self._sync_wrapper(fn, *args, **kwargs) [ 802.187589] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 802.187589] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] self.wait() [ 802.187589] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 802.187589] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] self[:] = self._gt.wait() [ 802.187589] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 802.187589] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] return self._exit_event.wait() [ 802.187589] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 802.187589] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] current.throw(*self._exc) [ 802.187965] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 802.187965] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] result = function(*args, **kwargs) [ 802.187965] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 802.187965] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] return func(*args, **kwargs) [ 802.187965] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 802.187965] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] raise e [ 802.187965] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 802.187965] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] nwinfo = self.network_api.allocate_for_instance( [ 802.187965] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 802.187965] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] created_port_ids = self._update_ports_for_instance( [ 802.187965] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 802.187965] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] with excutils.save_and_reraise_exception(): [ 802.187965] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 802.188297] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] self.force_reraise() [ 802.188297] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 802.188297] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] raise self.value [ 802.188297] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 802.188297] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] updated_port = self._update_port( [ 802.188297] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 802.188297] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] _ensure_no_port_binding_failure(port) [ 802.188297] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 802.188297] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] raise exception.PortBindingFailed(port_id=port['id']) [ 802.188297] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] nova.exception.PortBindingFailed: Binding failed for port 7434460f-bcc3-4312-8a89-15154d936f92, please check neutron logs for more information. [ 802.188297] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] [ 802.188297] env[61970]: INFO nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Terminating instance [ 802.189421] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquiring lock "refresh_cache-07eeb310-312b-4f32-bd5f-7b254cff0fe1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.189573] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquired lock "refresh_cache-07eeb310-312b-4f32-bd5f-7b254cff0fe1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.189741] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 802.218878] env[61970]: DEBUG oslo_concurrency.lockutils [req-18be19f5-75d0-49db-a675-4ab0ed7c4a87 req-611fd97b-43e9-46c3-9d73-28b3f2d3bcb2 service nova] Releasing lock "refresh_cache-3528d9b6-f3fb-43e9-be27-113d887a3414" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.219175] env[61970]: DEBUG nova.compute.manager [req-18be19f5-75d0-49db-a675-4ab0ed7c4a87 req-611fd97b-43e9-46c3-9d73-28b3f2d3bcb2 service nova] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Received event network-vif-deleted-c2e6332a-9fc1-4dbb-aea3-77565735c463 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.382806] env[61970]: DEBUG nova.compute.manager [req-f711d240-9fd0-4b94-bc00-bfe2a5ab034e req-7073be38-744a-4771-bc1f-abfda497830a service nova] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Received event network-changed-7434460f-bcc3-4312-8a89-15154d936f92 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.383036] env[61970]: DEBUG nova.compute.manager [req-f711d240-9fd0-4b94-bc00-bfe2a5ab034e req-7073be38-744a-4771-bc1f-abfda497830a service nova] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Refreshing instance network info cache due to event network-changed-7434460f-bcc3-4312-8a89-15154d936f92. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 802.383221] env[61970]: DEBUG oslo_concurrency.lockutils [req-f711d240-9fd0-4b94-bc00-bfe2a5ab034e req-7073be38-744a-4771-bc1f-abfda497830a service nova] Acquiring lock "refresh_cache-07eeb310-312b-4f32-bd5f-7b254cff0fe1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.644524] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.064s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.645053] env[61970]: DEBUG nova.compute.manager [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 802.647585] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.608s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.707675] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.799197] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.153199] env[61970]: DEBUG nova.compute.utils [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 803.158510] env[61970]: DEBUG nova.compute.manager [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 803.158713] env[61970]: DEBUG nova.network.neutron [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 803.197591] env[61970]: DEBUG nova.policy [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b45dc9d71c0e4693ba42322efde6c572', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5c64cb9232fb413cbd7627dcf077e9ef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 803.302172] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Releasing lock "refresh_cache-07eeb310-312b-4f32-bd5f-7b254cff0fe1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.302613] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 803.302793] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 803.303132] env[61970]: DEBUG oslo_concurrency.lockutils [req-f711d240-9fd0-4b94-bc00-bfe2a5ab034e req-7073be38-744a-4771-bc1f-abfda497830a service nova] Acquired lock "refresh_cache-07eeb310-312b-4f32-bd5f-7b254cff0fe1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.303387] env[61970]: DEBUG nova.network.neutron [req-f711d240-9fd0-4b94-bc00-bfe2a5ab034e req-7073be38-744a-4771-bc1f-abfda497830a service nova] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Refreshing network info cache for port 7434460f-bcc3-4312-8a89-15154d936f92 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 803.304376] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-831b33ba-0ba0-4950-acc4-ca88b1ff316f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.317395] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f12843-b27e-4a5e-bce0-fc2a2aa103fd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.340900] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 07eeb310-312b-4f32-bd5f-7b254cff0fe1 could not be found. [ 803.341135] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 803.341318] env[61970]: INFO nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Took 0.04 seconds to destroy the instance on the hypervisor. [ 803.341560] env[61970]: DEBUG oslo.service.loopingcall [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 803.343851] env[61970]: DEBUG nova.compute.manager [-] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 803.343952] env[61970]: DEBUG nova.network.neutron [-] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 803.372962] env[61970]: DEBUG nova.network.neutron [-] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 803.432948] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-353fc892-9adf-4410-96b3-a32c2ebf0044 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.440073] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72adc43c-ddbc-4156-9584-bd2de00753cf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.471608] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c9f916d-1b0c-4a9f-85b0-cd12805d4ad3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.477868] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eac77f4-70cf-44e9-99f3-b53ddaca1054 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.490919] env[61970]: DEBUG nova.compute.provider_tree [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.527323] env[61970]: DEBUG nova.network.neutron [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Successfully created port: 0ce24b8d-7846-42d7-817e-8118b9bff9ee {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 803.659000] env[61970]: DEBUG nova.compute.manager [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 803.834207] env[61970]: DEBUG nova.network.neutron [req-f711d240-9fd0-4b94-bc00-bfe2a5ab034e req-7073be38-744a-4771-bc1f-abfda497830a service nova] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 803.878624] env[61970]: DEBUG nova.network.neutron [-] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.000598] env[61970]: DEBUG nova.scheduler.client.report [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.103590] env[61970]: DEBUG nova.network.neutron [req-f711d240-9fd0-4b94-bc00-bfe2a5ab034e req-7073be38-744a-4771-bc1f-abfda497830a service nova] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.383071] env[61970]: INFO nova.compute.manager [-] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Took 1.04 seconds to deallocate network for instance. [ 804.385496] env[61970]: DEBUG nova.compute.claims [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 804.385660] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.437996] env[61970]: DEBUG nova.compute.manager [req-a5d3a4dd-5be6-4a0b-b1d2-46e1ec801435 req-c1e77d76-1a26-4eb7-b38f-bd772a170010 service nova] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Received event network-changed-0ce24b8d-7846-42d7-817e-8118b9bff9ee {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.438212] env[61970]: DEBUG nova.compute.manager [req-a5d3a4dd-5be6-4a0b-b1d2-46e1ec801435 req-c1e77d76-1a26-4eb7-b38f-bd772a170010 service nova] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Refreshing instance network info cache due to event network-changed-0ce24b8d-7846-42d7-817e-8118b9bff9ee. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 804.438428] env[61970]: DEBUG oslo_concurrency.lockutils [req-a5d3a4dd-5be6-4a0b-b1d2-46e1ec801435 req-c1e77d76-1a26-4eb7-b38f-bd772a170010 service nova] Acquiring lock "refresh_cache-a8ecafbe-f1e1-49f1-945d-ef8b6d15b709" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.438568] env[61970]: DEBUG oslo_concurrency.lockutils [req-a5d3a4dd-5be6-4a0b-b1d2-46e1ec801435 req-c1e77d76-1a26-4eb7-b38f-bd772a170010 service nova] Acquired lock "refresh_cache-a8ecafbe-f1e1-49f1-945d-ef8b6d15b709" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.438724] env[61970]: DEBUG nova.network.neutron [req-a5d3a4dd-5be6-4a0b-b1d2-46e1ec801435 req-c1e77d76-1a26-4eb7-b38f-bd772a170010 service nova] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Refreshing network info cache for port 0ce24b8d-7846-42d7-817e-8118b9bff9ee {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 804.508458] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.861s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.509096] env[61970]: ERROR nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e37d7d20-8753-4ad6-98b7-c7dee3b1c724, please check neutron logs for more information. [ 804.509096] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Traceback (most recent call last): [ 804.509096] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 804.509096] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] self.driver.spawn(context, instance, image_meta, [ 804.509096] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 804.509096] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 804.509096] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 804.509096] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] vm_ref = self.build_virtual_machine(instance, [ 804.509096] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 804.509096] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] vif_infos = vmwarevif.get_vif_info(self._session, [ 804.509096] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 804.509565] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] for vif in network_info: [ 804.509565] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 804.509565] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] return self._sync_wrapper(fn, *args, **kwargs) [ 804.509565] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 804.509565] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] self.wait() [ 804.509565] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 804.509565] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] self[:] = self._gt.wait() [ 804.509565] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 804.509565] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] return self._exit_event.wait() [ 804.509565] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 804.509565] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] result = hub.switch() [ 804.509565] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 804.509565] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] return self.greenlet.switch() [ 804.510106] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 804.510106] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] result = function(*args, **kwargs) [ 804.510106] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 804.510106] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] return func(*args, **kwargs) [ 804.510106] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 804.510106] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] raise e [ 804.510106] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 804.510106] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] nwinfo = self.network_api.allocate_for_instance( [ 804.510106] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.510106] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] created_port_ids = self._update_ports_for_instance( [ 804.510106] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.510106] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] with excutils.save_and_reraise_exception(): [ 804.510106] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.510656] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] self.force_reraise() [ 804.510656] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.510656] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] raise self.value [ 804.510656] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.510656] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] updated_port = self._update_port( [ 804.510656] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.510656] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] _ensure_no_port_binding_failure(port) [ 804.510656] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.510656] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] raise exception.PortBindingFailed(port_id=port['id']) [ 804.510656] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] nova.exception.PortBindingFailed: Binding failed for port e37d7d20-8753-4ad6-98b7-c7dee3b1c724, please check neutron logs for more information. [ 804.510656] env[61970]: ERROR nova.compute.manager [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] [ 804.511088] env[61970]: DEBUG nova.compute.utils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Binding failed for port e37d7d20-8753-4ad6-98b7-c7dee3b1c724, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 804.511088] env[61970]: DEBUG oslo_concurrency.lockutils [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.466s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.512492] env[61970]: INFO nova.compute.claims [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 804.515281] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Build of instance 01ee68e2-4310-477e-8adb-6ad67d6a65e7 was re-scheduled: Binding failed for port e37d7d20-8753-4ad6-98b7-c7dee3b1c724, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 804.515694] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 804.515910] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "refresh_cache-01ee68e2-4310-477e-8adb-6ad67d6a65e7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.516065] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquired lock "refresh_cache-01ee68e2-4310-477e-8adb-6ad67d6a65e7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.516222] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 804.604845] env[61970]: DEBUG oslo_concurrency.lockutils [req-f711d240-9fd0-4b94-bc00-bfe2a5ab034e req-7073be38-744a-4771-bc1f-abfda497830a service nova] Releasing lock "refresh_cache-07eeb310-312b-4f32-bd5f-7b254cff0fe1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.605113] env[61970]: DEBUG nova.compute.manager [req-f711d240-9fd0-4b94-bc00-bfe2a5ab034e req-7073be38-744a-4771-bc1f-abfda497830a service nova] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Received event network-vif-deleted-7434460f-bcc3-4312-8a89-15154d936f92 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.627498] env[61970]: ERROR nova.compute.manager [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0ce24b8d-7846-42d7-817e-8118b9bff9ee, please check neutron logs for more information. [ 804.627498] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 804.627498] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 804.627498] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 804.627498] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.627498] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 804.627498] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.627498] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 804.627498] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.627498] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 804.627498] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.627498] env[61970]: ERROR nova.compute.manager raise self.value [ 804.627498] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.627498] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 804.627498] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.627498] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 804.627996] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.627996] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 804.627996] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0ce24b8d-7846-42d7-817e-8118b9bff9ee, please check neutron logs for more information. [ 804.627996] env[61970]: ERROR nova.compute.manager [ 804.627996] env[61970]: Traceback (most recent call last): [ 804.627996] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 804.627996] env[61970]: listener.cb(fileno) [ 804.627996] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 804.627996] env[61970]: result = function(*args, **kwargs) [ 804.627996] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 804.627996] env[61970]: return func(*args, **kwargs) [ 804.627996] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 804.627996] env[61970]: raise e [ 804.627996] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 804.627996] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 804.627996] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.627996] env[61970]: created_port_ids = self._update_ports_for_instance( [ 804.627996] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.627996] env[61970]: with excutils.save_and_reraise_exception(): [ 804.627996] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.627996] env[61970]: self.force_reraise() [ 804.627996] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.627996] env[61970]: raise self.value [ 804.627996] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.627996] env[61970]: updated_port = self._update_port( [ 804.627996] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.627996] env[61970]: _ensure_no_port_binding_failure(port) [ 804.627996] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.627996] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 804.628772] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 0ce24b8d-7846-42d7-817e-8118b9bff9ee, please check neutron logs for more information. [ 804.628772] env[61970]: Removing descriptor: 15 [ 804.672843] env[61970]: DEBUG nova.compute.manager [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 804.701539] env[61970]: DEBUG nova.virt.hardware [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 804.701778] env[61970]: DEBUG nova.virt.hardware [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 804.702223] env[61970]: DEBUG nova.virt.hardware [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 804.702223] env[61970]: DEBUG nova.virt.hardware [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 804.702330] env[61970]: DEBUG nova.virt.hardware [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 804.702425] env[61970]: DEBUG nova.virt.hardware [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 804.702629] env[61970]: DEBUG nova.virt.hardware [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 804.702820] env[61970]: DEBUG nova.virt.hardware [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 804.702942] env[61970]: DEBUG nova.virt.hardware [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 804.703130] env[61970]: DEBUG nova.virt.hardware [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 804.703319] env[61970]: DEBUG nova.virt.hardware [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 804.704230] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8a604df-2eee-4ee5-b801-4599b38c781f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.712509] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d4d65cf-6335-40d3-ab6f-257b84c34031 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.725721] env[61970]: ERROR nova.compute.manager [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0ce24b8d-7846-42d7-817e-8118b9bff9ee, please check neutron logs for more information. [ 804.725721] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Traceback (most recent call last): [ 804.725721] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 804.725721] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] yield resources [ 804.725721] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 804.725721] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] self.driver.spawn(context, instance, image_meta, [ 804.725721] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 804.725721] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] self._vmops.spawn(context, instance, image_meta, injected_files, [ 804.725721] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 804.725721] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] vm_ref = self.build_virtual_machine(instance, [ 804.725721] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 804.726060] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] vif_infos = vmwarevif.get_vif_info(self._session, [ 804.726060] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 804.726060] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] for vif in network_info: [ 804.726060] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 804.726060] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] return self._sync_wrapper(fn, *args, **kwargs) [ 804.726060] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 804.726060] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] self.wait() [ 804.726060] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 804.726060] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] self[:] = self._gt.wait() [ 804.726060] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 804.726060] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] return self._exit_event.wait() [ 804.726060] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 804.726060] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] current.throw(*self._exc) [ 804.726492] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 804.726492] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] result = function(*args, **kwargs) [ 804.726492] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 804.726492] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] return func(*args, **kwargs) [ 804.726492] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 804.726492] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] raise e [ 804.726492] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 804.726492] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] nwinfo = self.network_api.allocate_for_instance( [ 804.726492] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.726492] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] created_port_ids = self._update_ports_for_instance( [ 804.726492] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.726492] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] with excutils.save_and_reraise_exception(): [ 804.726492] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.726844] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] self.force_reraise() [ 804.726844] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.726844] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] raise self.value [ 804.726844] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.726844] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] updated_port = self._update_port( [ 804.726844] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.726844] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] _ensure_no_port_binding_failure(port) [ 804.726844] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.726844] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] raise exception.PortBindingFailed(port_id=port['id']) [ 804.726844] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] nova.exception.PortBindingFailed: Binding failed for port 0ce24b8d-7846-42d7-817e-8118b9bff9ee, please check neutron logs for more information. [ 804.726844] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] [ 804.726844] env[61970]: INFO nova.compute.manager [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Terminating instance [ 804.728135] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "refresh_cache-a8ecafbe-f1e1-49f1-945d-ef8b6d15b709" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.959819] env[61970]: DEBUG nova.network.neutron [req-a5d3a4dd-5be6-4a0b-b1d2-46e1ec801435 req-c1e77d76-1a26-4eb7-b38f-bd772a170010 service nova] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.029619] env[61970]: DEBUG nova.network.neutron [req-a5d3a4dd-5be6-4a0b-b1d2-46e1ec801435 req-c1e77d76-1a26-4eb7-b38f-bd772a170010 service nova] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.039598] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.108237] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.532367] env[61970]: DEBUG oslo_concurrency.lockutils [req-a5d3a4dd-5be6-4a0b-b1d2-46e1ec801435 req-c1e77d76-1a26-4eb7-b38f-bd772a170010 service nova] Releasing lock "refresh_cache-a8ecafbe-f1e1-49f1-945d-ef8b6d15b709" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.532913] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "refresh_cache-a8ecafbe-f1e1-49f1-945d-ef8b6d15b709" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.533172] env[61970]: DEBUG nova.network.neutron [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 805.610798] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Releasing lock "refresh_cache-01ee68e2-4310-477e-8adb-6ad67d6a65e7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.611032] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 805.611215] env[61970]: DEBUG nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 805.611395] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 805.627681] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.774341] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c5c6e9-c1bb-4378-ad19-83bdb06692a3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.781970] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99aa92bc-1d94-4383-8bcb-66b02f951c3e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.810322] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-291eb480-8650-44b4-a92e-f971cec64a7a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.817585] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ff46a76-305c-468b-bcd1-3449fd347935 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.830290] env[61970]: DEBUG nova.compute.provider_tree [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.050783] env[61970]: DEBUG nova.network.neutron [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.130258] env[61970]: DEBUG nova.network.neutron [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.138285] env[61970]: DEBUG nova.network.neutron [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.333094] env[61970]: DEBUG nova.scheduler.client.report [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.472067] env[61970]: DEBUG nova.compute.manager [req-8e18f335-5450-4d35-98a1-8b8f9d416fbd req-0bf8024c-3614-4051-8d11-a9ecfd9eb886 service nova] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Received event network-vif-deleted-0ce24b8d-7846-42d7-817e-8118b9bff9ee {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.632733] env[61970]: INFO nova.compute.manager [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 01ee68e2-4310-477e-8adb-6ad67d6a65e7] Took 1.02 seconds to deallocate network for instance. [ 806.641012] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "refresh_cache-a8ecafbe-f1e1-49f1-945d-ef8b6d15b709" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.641414] env[61970]: DEBUG nova.compute.manager [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 806.641623] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 806.641903] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-30ae1503-12e9-4aa2-b866-7d0c28f01566 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.652058] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c9c9812-580e-4343-83d7-76af282af42b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.673493] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a8ecafbe-f1e1-49f1-945d-ef8b6d15b709 could not be found. [ 806.673672] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 806.673851] env[61970]: INFO nova.compute.manager [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Took 0.03 seconds to destroy the instance on the hypervisor. [ 806.674096] env[61970]: DEBUG oslo.service.loopingcall [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.674928] env[61970]: DEBUG nova.compute.manager [-] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 806.675038] env[61970]: DEBUG nova.network.neutron [-] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 806.689775] env[61970]: DEBUG nova.network.neutron [-] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.838693] env[61970]: DEBUG oslo_concurrency.lockutils [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.328s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.839232] env[61970]: DEBUG nova.compute.manager [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 806.841831] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.974s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.191763] env[61970]: DEBUG nova.network.neutron [-] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.346166] env[61970]: DEBUG nova.compute.utils [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 807.350870] env[61970]: DEBUG nova.compute.manager [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Not allocating networking since 'none' was specified. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 807.648392] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4376ef7d-cf33-4783-bb37-9a91f267ffb1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.655860] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d4a02f-d04f-4d2a-a892-a409cc88765d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.659340] env[61970]: INFO nova.scheduler.client.report [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Deleted allocations for instance 01ee68e2-4310-477e-8adb-6ad67d6a65e7 [ 807.693744] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e3b806-99d0-4564-bd4e-d80e1d83e863 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.696409] env[61970]: INFO nova.compute.manager [-] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Took 1.02 seconds to deallocate network for instance. [ 807.702410] env[61970]: DEBUG nova.compute.claims [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 807.702556] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.703718] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22cb8bd3-dd17-423e-aff1-ceb53ecb8c0d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.717337] env[61970]: DEBUG nova.compute.provider_tree [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.851797] env[61970]: DEBUG nova.compute.manager [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 808.167873] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6cf08d3d-955a-4f30-9d04-009d82c99b33 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "01ee68e2-4310-477e-8adb-6ad67d6a65e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.732s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.220874] env[61970]: DEBUG nova.scheduler.client.report [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.672758] env[61970]: DEBUG nova.compute.manager [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 808.725487] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.884s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.726131] env[61970]: ERROR nova.compute.manager [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5b0c073f-e9de-4f25-ab23-376446d9aad6, please check neutron logs for more information. [ 808.726131] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] Traceback (most recent call last): [ 808.726131] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 808.726131] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] self.driver.spawn(context, instance, image_meta, [ 808.726131] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 808.726131] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] self._vmops.spawn(context, instance, image_meta, injected_files, [ 808.726131] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 808.726131] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] vm_ref = self.build_virtual_machine(instance, [ 808.726131] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 808.726131] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] vif_infos = vmwarevif.get_vif_info(self._session, [ 808.726131] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 808.726444] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] for vif in network_info: [ 808.726444] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 808.726444] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] return self._sync_wrapper(fn, *args, **kwargs) [ 808.726444] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 808.726444] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] self.wait() [ 808.726444] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 808.726444] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] self[:] = self._gt.wait() [ 808.726444] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 808.726444] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] return self._exit_event.wait() [ 808.726444] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 808.726444] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] current.throw(*self._exc) [ 808.726444] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 808.726444] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] result = function(*args, **kwargs) [ 808.726763] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 808.726763] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] return func(*args, **kwargs) [ 808.726763] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 808.726763] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] raise e [ 808.726763] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 808.726763] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] nwinfo = self.network_api.allocate_for_instance( [ 808.726763] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 808.726763] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] created_port_ids = self._update_ports_for_instance( [ 808.726763] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 808.726763] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] with excutils.save_and_reraise_exception(): [ 808.726763] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 808.726763] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] self.force_reraise() [ 808.726763] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 808.727118] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] raise self.value [ 808.727118] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 808.727118] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] updated_port = self._update_port( [ 808.727118] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 808.727118] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] _ensure_no_port_binding_failure(port) [ 808.727118] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 808.727118] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] raise exception.PortBindingFailed(port_id=port['id']) [ 808.727118] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] nova.exception.PortBindingFailed: Binding failed for port 5b0c073f-e9de-4f25-ab23-376446d9aad6, please check neutron logs for more information. [ 808.727118] env[61970]: ERROR nova.compute.manager [instance: 95d428e6-fd95-4628-9053-76d218f39258] [ 808.727118] env[61970]: DEBUG nova.compute.utils [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Binding failed for port 5b0c073f-e9de-4f25-ab23-376446d9aad6, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 808.728011] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.418s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.729422] env[61970]: INFO nova.compute.claims [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 808.732026] env[61970]: DEBUG nova.compute.manager [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Build of instance 95d428e6-fd95-4628-9053-76d218f39258 was re-scheduled: Binding failed for port 5b0c073f-e9de-4f25-ab23-376446d9aad6, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 808.732463] env[61970]: DEBUG nova.compute.manager [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 808.732696] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Acquiring lock "refresh_cache-95d428e6-fd95-4628-9053-76d218f39258" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.732839] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Acquired lock "refresh_cache-95d428e6-fd95-4628-9053-76d218f39258" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.732995] env[61970]: DEBUG nova.network.neutron [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 808.861467] env[61970]: DEBUG nova.compute.manager [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 808.887724] env[61970]: DEBUG nova.virt.hardware [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 808.887974] env[61970]: DEBUG nova.virt.hardware [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 808.888150] env[61970]: DEBUG nova.virt.hardware [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 808.888335] env[61970]: DEBUG nova.virt.hardware [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 808.888481] env[61970]: DEBUG nova.virt.hardware [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 808.888659] env[61970]: DEBUG nova.virt.hardware [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 808.888825] env[61970]: DEBUG nova.virt.hardware [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 808.888982] env[61970]: DEBUG nova.virt.hardware [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 808.889164] env[61970]: DEBUG nova.virt.hardware [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 808.889327] env[61970]: DEBUG nova.virt.hardware [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 808.889980] env[61970]: DEBUG nova.virt.hardware [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 808.891025] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25be0bd0-4554-4485-aea7-fb6e1e022be5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.899116] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87c61acd-3a6a-4d05-a72d-ef17501c180a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.913636] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Instance VIF info [] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 808.919212] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Creating folder: Project (7e62049494dd4e6cb3bc62cd6afe6138). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 808.919525] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-37f807a0-1a6c-469c-8512-12efa0836ed0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.932840] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Created folder: Project (7e62049494dd4e6cb3bc62cd6afe6138) in parent group-v288740. [ 808.933075] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Creating folder: Instances. Parent ref: group-v288765. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 808.933266] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03908b45-aa3e-4819-8fb5-05b16ad966ce {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.942666] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Created folder: Instances in parent group-v288765. [ 808.942900] env[61970]: DEBUG oslo.service.loopingcall [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.943099] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 808.943307] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-58c4ad97-1d6d-4c1d-ac52-6acae082384d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.959374] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 808.959374] env[61970]: value = "task-1355560" [ 808.959374] env[61970]: _type = "Task" [ 808.959374] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.966626] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355560, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.194775] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.255410] env[61970]: DEBUG nova.network.neutron [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.371195] env[61970]: DEBUG nova.network.neutron [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.469011] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355560, 'name': CreateVM_Task, 'duration_secs': 0.246408} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.469195] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 809.469620] env[61970]: DEBUG oslo_concurrency.lockutils [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.469780] env[61970]: DEBUG oslo_concurrency.lockutils [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.470104] env[61970]: DEBUG oslo_concurrency.lockutils [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 809.470401] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d7c5b36-d932-43d3-b04c-2754b5ab2e32 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.474971] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 809.474971] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5210c39d-0e09-a3c9-25ca-73f8c4b7d099" [ 809.474971] env[61970]: _type = "Task" [ 809.474971] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.482275] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5210c39d-0e09-a3c9-25ca-73f8c4b7d099, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.873500] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Releasing lock "refresh_cache-95d428e6-fd95-4628-9053-76d218f39258" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.873929] env[61970]: DEBUG nova.compute.manager [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 809.874153] env[61970]: DEBUG nova.compute.manager [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 809.874323] env[61970]: DEBUG nova.network.neutron [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 809.889448] env[61970]: DEBUG nova.network.neutron [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.987140] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5210c39d-0e09-a3c9-25ca-73f8c4b7d099, 'name': SearchDatastore_Task, 'duration_secs': 0.010523} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.987457] env[61970]: DEBUG oslo_concurrency.lockutils [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.987712] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.987939] env[61970]: DEBUG oslo_concurrency.lockutils [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.988098] env[61970]: DEBUG oslo_concurrency.lockutils [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.988284] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.988540] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a63fccce-79ae-4d10-8ed5-b50aa3c583d4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.991547] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c97acfe-1415-4a0e-8c52-67f88ce7420c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.998123] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628f1b4d-670e-4764-a6b6-87423e9950d7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.002242] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 810.002393] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 810.003406] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e42c24b4-253a-4d11-9220-a5596853db9c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.032696] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dff1a95b-41d6-4a39-b8d6-6334be2742ca {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.035820] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 810.035820] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5269fa11-99bc-00c7-3531-06ce025229fa" [ 810.035820] env[61970]: _type = "Task" [ 810.035820] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.042600] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8457477e-6b4d-4084-b26d-dc1fd69baafb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.049444] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5269fa11-99bc-00c7-3531-06ce025229fa, 'name': SearchDatastore_Task, 'duration_secs': 0.009787} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.050546] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b776b61-3faf-406e-b9f5-dbd463b1f612 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.060154] env[61970]: DEBUG nova.compute.provider_tree [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.067021] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 810.067021] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d9a406-8309-fd92-55f0-393362ede574" [ 810.067021] env[61970]: _type = "Task" [ 810.067021] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.073834] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d9a406-8309-fd92-55f0-393362ede574, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.095316] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "99200b46-7eb7-4ca2-b352-b91b4ad076af" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.095559] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "99200b46-7eb7-4ca2-b352-b91b4ad076af" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.120687] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "e2185ed0-2bc1-4718-b47d-57150b5e60ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.120902] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "e2185ed0-2bc1-4718-b47d-57150b5e60ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.391834] env[61970]: DEBUG nova.network.neutron [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.563543] env[61970]: DEBUG nova.scheduler.client.report [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 810.575650] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d9a406-8309-fd92-55f0-393362ede574, 'name': SearchDatastore_Task, 'duration_secs': 0.010036} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.575878] env[61970]: DEBUG oslo_concurrency.lockutils [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.577025] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0/7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 810.577025] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f238ca74-ccfb-443b-a267-a94f3adbb01e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.584080] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 810.584080] env[61970]: value = "task-1355561" [ 810.584080] env[61970]: _type = "Task" [ 810.584080] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.591592] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355561, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.894468] env[61970]: INFO nova.compute.manager [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] [instance: 95d428e6-fd95-4628-9053-76d218f39258] Took 1.02 seconds to deallocate network for instance. [ 811.071499] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.343s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.072017] env[61970]: DEBUG nova.compute.manager [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 811.074539] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.070s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.093141] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355561, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451249} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.093451] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0/7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 811.093638] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 811.093898] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9af30b8d-c10a-4a2b-9ae5-2edbc41d03bc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.101531] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 811.101531] env[61970]: value = "task-1355562" [ 811.101531] env[61970]: _type = "Task" [ 811.101531] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.112201] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355562, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.579108] env[61970]: DEBUG nova.compute.utils [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 811.583168] env[61970]: DEBUG nova.compute.manager [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Not allocating networking since 'none' was specified. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 811.610209] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355562, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060621} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.612501] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 811.613454] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-895fb819-b2d1-4bf6-ba9a-33637176b678 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.632537] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0/7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 811.634880] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-79122de0-49e5-4b85-9144-7ffcec74653e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.654970] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 811.654970] env[61970]: value = "task-1355563" [ 811.654970] env[61970]: _type = "Task" [ 811.654970] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.664854] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355563, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.861493] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d931880-d36b-4424-815c-550b459fa37e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.869191] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e578d729-3c29-41f8-a8f5-4bb54e15f62d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.899240] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54bb14a9-7909-4cd6-8ee1-60f89c6a258c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.908696] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06351048-c567-4a65-96a4-3a326d2e61d7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.921624] env[61970]: DEBUG nova.compute.provider_tree [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.924059] env[61970]: INFO nova.scheduler.client.report [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Deleted allocations for instance 95d428e6-fd95-4628-9053-76d218f39258 [ 812.086044] env[61970]: DEBUG nova.compute.manager [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 812.164676] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355563, 'name': ReconfigVM_Task, 'duration_secs': 0.27386} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.164944] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0/7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 812.165566] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-029d92ba-0688-434c-b7b6-ed8350ed5f62 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.171210] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 812.171210] env[61970]: value = "task-1355564" [ 812.171210] env[61970]: _type = "Task" [ 812.171210] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.179408] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355564, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.431044] env[61970]: DEBUG nova.scheduler.client.report [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 812.433550] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85c09818-cbd1-48e9-8e40-490b029511dd tempest-SecurityGroupsTestJSON-1798428103 tempest-SecurityGroupsTestJSON-1798428103-project-member] Lock "95d428e6-fd95-4628-9053-76d218f39258" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.993s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.680635] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355564, 'name': Rename_Task, 'duration_secs': 0.154384} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.680906] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 812.681620] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f24048a4-8a8d-4c2b-9637-90559f4f1f59 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.687964] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 812.687964] env[61970]: value = "task-1355565" [ 812.687964] env[61970]: _type = "Task" [ 812.687964] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.695660] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355565, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.935086] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.860s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.935843] env[61970]: ERROR nova.compute.manager [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f8eba464-80eb-4d70-af1e-f0a5fc33895b, please check neutron logs for more information. [ 812.935843] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Traceback (most recent call last): [ 812.935843] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 812.935843] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] self.driver.spawn(context, instance, image_meta, [ 812.935843] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 812.935843] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] self._vmops.spawn(context, instance, image_meta, injected_files, [ 812.935843] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 812.935843] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] vm_ref = self.build_virtual_machine(instance, [ 812.935843] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 812.935843] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] vif_infos = vmwarevif.get_vif_info(self._session, [ 812.935843] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 812.936209] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] for vif in network_info: [ 812.936209] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 812.936209] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] return self._sync_wrapper(fn, *args, **kwargs) [ 812.936209] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 812.936209] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] self.wait() [ 812.936209] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 812.936209] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] self[:] = self._gt.wait() [ 812.936209] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 812.936209] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] return self._exit_event.wait() [ 812.936209] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 812.936209] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] result = hub.switch() [ 812.936209] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 812.936209] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] return self.greenlet.switch() [ 812.936610] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 812.936610] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] result = function(*args, **kwargs) [ 812.936610] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 812.936610] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] return func(*args, **kwargs) [ 812.936610] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 812.936610] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] raise e [ 812.936610] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 812.936610] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] nwinfo = self.network_api.allocate_for_instance( [ 812.936610] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 812.936610] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] created_port_ids = self._update_ports_for_instance( [ 812.936610] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 812.936610] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] with excutils.save_and_reraise_exception(): [ 812.936610] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 812.937104] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] self.force_reraise() [ 812.937104] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 812.937104] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] raise self.value [ 812.937104] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 812.937104] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] updated_port = self._update_port( [ 812.937104] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 812.937104] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] _ensure_no_port_binding_failure(port) [ 812.937104] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 812.937104] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] raise exception.PortBindingFailed(port_id=port['id']) [ 812.937104] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] nova.exception.PortBindingFailed: Binding failed for port f8eba464-80eb-4d70-af1e-f0a5fc33895b, please check neutron logs for more information. [ 812.937104] env[61970]: ERROR nova.compute.manager [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] [ 812.937444] env[61970]: DEBUG nova.compute.utils [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Binding failed for port f8eba464-80eb-4d70-af1e-f0a5fc33895b, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 812.938024] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.795s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.939679] env[61970]: INFO nova.compute.claims [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 812.942369] env[61970]: DEBUG nova.compute.manager [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 812.945267] env[61970]: DEBUG nova.compute.manager [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Build of instance 3f0767a0-9925-4cb9-8032-b5ca071cd572 was re-scheduled: Binding failed for port f8eba464-80eb-4d70-af1e-f0a5fc33895b, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 812.946962] env[61970]: DEBUG nova.compute.manager [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 812.947973] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Acquiring lock "refresh_cache-3f0767a0-9925-4cb9-8032-b5ca071cd572" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.947973] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Acquired lock "refresh_cache-3f0767a0-9925-4cb9-8032-b5ca071cd572" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.947973] env[61970]: DEBUG nova.network.neutron [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 813.095074] env[61970]: DEBUG nova.compute.manager [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 813.122248] env[61970]: DEBUG nova.virt.hardware [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 813.122564] env[61970]: DEBUG nova.virt.hardware [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 813.122730] env[61970]: DEBUG nova.virt.hardware [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 813.122914] env[61970]: DEBUG nova.virt.hardware [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 813.123069] env[61970]: DEBUG nova.virt.hardware [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 813.123216] env[61970]: DEBUG nova.virt.hardware [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 813.123427] env[61970]: DEBUG nova.virt.hardware [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 813.123640] env[61970]: DEBUG nova.virt.hardware [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 813.123824] env[61970]: DEBUG nova.virt.hardware [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 813.123986] env[61970]: DEBUG nova.virt.hardware [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 813.124169] env[61970]: DEBUG nova.virt.hardware [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 813.125048] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-541be346-cce7-44d6-8fd4-7d8d380494fb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.132903] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458fb2c0-4f0b-41ab-af0f-94f4bfa9d8f4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.148026] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Instance VIF info [] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 813.154038] env[61970]: DEBUG oslo.service.loopingcall [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 813.154277] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 813.154479] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-95fea0b4-4469-4701-b7e6-321ad1d64064 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.172859] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 813.172859] env[61970]: value = "task-1355566" [ 813.172859] env[61970]: _type = "Task" [ 813.172859] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.180423] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355566, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.196530] env[61970]: DEBUG oslo_vmware.api [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355565, 'name': PowerOnVM_Task, 'duration_secs': 0.407678} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.199424] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 813.199424] env[61970]: INFO nova.compute.manager [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Took 4.34 seconds to spawn the instance on the hypervisor. [ 813.199424] env[61970]: DEBUG nova.compute.manager [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 813.199424] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d34f4b5-ddf0-4975-9233-4fb728dee034 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.469551] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.470359] env[61970]: DEBUG nova.network.neutron [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 813.560862] env[61970]: DEBUG nova.network.neutron [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.682886] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355566, 'name': CreateVM_Task, 'duration_secs': 0.263187} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.683097] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 813.683554] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.683733] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.684056] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 813.684313] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-312b7280-ac8b-4150-8aea-a413b5f17735 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.688895] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 813.688895] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52088346-c423-206b-146a-4a2b552a4d67" [ 813.688895] env[61970]: _type = "Task" [ 813.688895] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.696522] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52088346-c423-206b-146a-4a2b552a4d67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.716860] env[61970]: INFO nova.compute.manager [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Took 36.69 seconds to build instance. [ 814.069716] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Releasing lock "refresh_cache-3f0767a0-9925-4cb9-8032-b5ca071cd572" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.069716] env[61970]: DEBUG nova.compute.manager [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 814.069716] env[61970]: DEBUG nova.compute.manager [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 814.069716] env[61970]: DEBUG nova.network.neutron [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 814.088893] env[61970]: DEBUG nova.network.neutron [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.207679] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52088346-c423-206b-146a-4a2b552a4d67, 'name': SearchDatastore_Task, 'duration_secs': 0.009752} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.211940] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.212348] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 814.212712] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.212953] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.213251] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 814.214265] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-144b1634-1515-4629-9d93-270049c4c19b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.218369] env[61970]: DEBUG oslo_concurrency.lockutils [None req-910e029a-71ab-4ade-8837-9c9373bab30f tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.312s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.225726] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 814.225926] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 814.229259] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ad1f5a4-f348-46f2-a1eb-d7d793ed5026 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.236823] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 814.236823] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52dd949b-15a1-a2e0-848a-0408c967b3e7" [ 814.236823] env[61970]: _type = "Task" [ 814.236823] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.244345] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52dd949b-15a1-a2e0-848a-0408c967b3e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.355394] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d8b6245-d5b4-4cd4-8139-e2b3a666b617 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.363730] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea613395-b8a2-4fe1-aea4-0e5b7952c73d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.395558] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-092db7a2-ef06-412e-9d4e-c491e44a42e5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.403869] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51492968-6968-4dbd-8cbe-deb2acd69adc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.422027] env[61970]: DEBUG nova.compute.provider_tree [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 814.592535] env[61970]: DEBUG nova.network.neutron [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.722428] env[61970]: DEBUG nova.compute.manager [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 814.745676] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52dd949b-15a1-a2e0-848a-0408c967b3e7, 'name': SearchDatastore_Task, 'duration_secs': 0.018115} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.746678] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a89b5927-976a-4fe8-804b-e203c4106b1c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.754020] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 814.754020] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]527dc6d5-bac3-47ad-fc6a-a30bbf912a20" [ 814.754020] env[61970]: _type = "Task" [ 814.754020] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.763235] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]527dc6d5-bac3-47ad-fc6a-a30bbf912a20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.923272] env[61970]: DEBUG nova.scheduler.client.report [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.095503] env[61970]: INFO nova.compute.manager [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] [instance: 3f0767a0-9925-4cb9-8032-b5ca071cd572] Took 1.03 seconds to deallocate network for instance. [ 815.246199] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.262356] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]527dc6d5-bac3-47ad-fc6a-a30bbf912a20, 'name': SearchDatastore_Task, 'duration_secs': 0.009977} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.262600] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.263069] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] a78b63a5-6bb8-4271-90d8-1e86fb29db4f/a78b63a5-6bb8-4271-90d8-1e86fb29db4f.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 815.263303] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4a455897-a1ab-401c-a404-ca9140744fff {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.272163] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 815.272163] env[61970]: value = "task-1355567" [ 815.272163] env[61970]: _type = "Task" [ 815.272163] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.280228] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355567, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.428582] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.490s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.429141] env[61970]: DEBUG nova.compute.manager [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 815.431872] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.807s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.781956] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355567, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487788} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.782239] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] a78b63a5-6bb8-4271-90d8-1e86fb29db4f/a78b63a5-6bb8-4271-90d8-1e86fb29db4f.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 815.782454] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 815.782742] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dcf6811f-f7ef-46f4-a72d-c94c5c4054bb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.788486] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 815.788486] env[61970]: value = "task-1355568" [ 815.788486] env[61970]: _type = "Task" [ 815.788486] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.795802] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355568, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.936292] env[61970]: DEBUG nova.compute.utils [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 815.941053] env[61970]: DEBUG nova.compute.manager [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 815.941053] env[61970]: DEBUG nova.network.neutron [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 816.004217] env[61970]: DEBUG nova.policy [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '33e981eb8a074d9bb9ad1db5b4136ec6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7921d2d3633e4562ae9689a5aece0662', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 816.127490] env[61970]: INFO nova.scheduler.client.report [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Deleted allocations for instance 3f0767a0-9925-4cb9-8032-b5ca071cd572 [ 816.237023] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08f35c6e-1ec0-42eb-b16f-9708a342e979 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.241178] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-570feebf-6328-44eb-87fc-700a4d5d7659 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.272632] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3104297c-3d42-4cc2-add9-b9bce3b89360 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.279875] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d86821-de8f-4c29-b4aa-6bf59b34fb6b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.293796] env[61970]: DEBUG nova.compute.provider_tree [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.295725] env[61970]: DEBUG nova.network.neutron [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Successfully created port: 9248c5b1-6dfe-4447-a7e4-805e208e2ce3 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 816.305414] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355568, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062513} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.306231] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 816.306992] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea42bd4-fb2e-4b24-8d68-5bd613b0c197 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.326757] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] a78b63a5-6bb8-4271-90d8-1e86fb29db4f/a78b63a5-6bb8-4271-90d8-1e86fb29db4f.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 816.327315] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2ee140e-95cf-4a73-a5bc-65973a801c9b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.346732] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 816.346732] env[61970]: value = "task-1355569" [ 816.346732] env[61970]: _type = "Task" [ 816.346732] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.354756] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355569, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.441707] env[61970]: DEBUG nova.compute.manager [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 816.637756] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8c4c8812-4017-4373-a167-951d9c61a785 tempest-VolumesAdminNegativeTest-155105357 tempest-VolumesAdminNegativeTest-155105357-project-member] Lock "3f0767a0-9925-4cb9-8032-b5ca071cd572" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.491s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.802207] env[61970]: DEBUG nova.scheduler.client.report [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.856925] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355569, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.059241] env[61970]: DEBUG nova.compute.manager [req-90e0ba58-e434-42e8-87c5-a618a82b60c2 req-ffdc6eff-3db4-4d0d-b634-ec0c15c0f04b service nova] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Received event network-changed-9248c5b1-6dfe-4447-a7e4-805e208e2ce3 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.059241] env[61970]: DEBUG nova.compute.manager [req-90e0ba58-e434-42e8-87c5-a618a82b60c2 req-ffdc6eff-3db4-4d0d-b634-ec0c15c0f04b service nova] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Refreshing instance network info cache due to event network-changed-9248c5b1-6dfe-4447-a7e4-805e208e2ce3. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 817.059412] env[61970]: DEBUG oslo_concurrency.lockutils [req-90e0ba58-e434-42e8-87c5-a618a82b60c2 req-ffdc6eff-3db4-4d0d-b634-ec0c15c0f04b service nova] Acquiring lock "refresh_cache-27996960-d2d4-496c-b52d-5d7312aed040" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.059461] env[61970]: DEBUG oslo_concurrency.lockutils [req-90e0ba58-e434-42e8-87c5-a618a82b60c2 req-ffdc6eff-3db4-4d0d-b634-ec0c15c0f04b service nova] Acquired lock "refresh_cache-27996960-d2d4-496c-b52d-5d7312aed040" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.059630] env[61970]: DEBUG nova.network.neutron [req-90e0ba58-e434-42e8-87c5-a618a82b60c2 req-ffdc6eff-3db4-4d0d-b634-ec0c15c0f04b service nova] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Refreshing network info cache for port 9248c5b1-6dfe-4447-a7e4-805e208e2ce3 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 817.141061] env[61970]: DEBUG nova.compute.manager [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 817.235571] env[61970]: ERROR nova.compute.manager [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9248c5b1-6dfe-4447-a7e4-805e208e2ce3, please check neutron logs for more information. [ 817.235571] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 817.235571] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 817.235571] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 817.235571] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 817.235571] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 817.235571] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 817.235571] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 817.235571] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 817.235571] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 817.235571] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 817.235571] env[61970]: ERROR nova.compute.manager raise self.value [ 817.235571] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 817.235571] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 817.235571] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 817.235571] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 817.236111] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 817.236111] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 817.236111] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9248c5b1-6dfe-4447-a7e4-805e208e2ce3, please check neutron logs for more information. [ 817.236111] env[61970]: ERROR nova.compute.manager [ 817.236111] env[61970]: Traceback (most recent call last): [ 817.236111] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 817.236111] env[61970]: listener.cb(fileno) [ 817.236111] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 817.236111] env[61970]: result = function(*args, **kwargs) [ 817.236111] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 817.236111] env[61970]: return func(*args, **kwargs) [ 817.236111] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 817.236111] env[61970]: raise e [ 817.236111] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 817.236111] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 817.236111] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 817.236111] env[61970]: created_port_ids = self._update_ports_for_instance( [ 817.236111] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 817.236111] env[61970]: with excutils.save_and_reraise_exception(): [ 817.236111] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 817.236111] env[61970]: self.force_reraise() [ 817.236111] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 817.236111] env[61970]: raise self.value [ 817.236111] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 817.236111] env[61970]: updated_port = self._update_port( [ 817.236111] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 817.236111] env[61970]: _ensure_no_port_binding_failure(port) [ 817.236111] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 817.236111] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 817.236854] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 9248c5b1-6dfe-4447-a7e4-805e208e2ce3, please check neutron logs for more information. [ 817.236854] env[61970]: Removing descriptor: 17 [ 817.307832] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.876s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.309079] env[61970]: ERROR nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3a2c383f-766e-454a-bddf-a884d879da19, please check neutron logs for more information. [ 817.309079] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Traceback (most recent call last): [ 817.309079] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 817.309079] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] self.driver.spawn(context, instance, image_meta, [ 817.309079] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 817.309079] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 817.309079] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 817.309079] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] vm_ref = self.build_virtual_machine(instance, [ 817.309079] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 817.309079] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] vif_infos = vmwarevif.get_vif_info(self._session, [ 817.309079] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 817.309428] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] for vif in network_info: [ 817.309428] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 817.309428] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] return self._sync_wrapper(fn, *args, **kwargs) [ 817.309428] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 817.309428] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] self.wait() [ 817.309428] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 817.309428] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] self[:] = self._gt.wait() [ 817.309428] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 817.309428] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] return self._exit_event.wait() [ 817.309428] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 817.309428] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] current.throw(*self._exc) [ 817.309428] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 817.309428] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] result = function(*args, **kwargs) [ 817.309752] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 817.309752] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] return func(*args, **kwargs) [ 817.309752] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 817.309752] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] raise e [ 817.309752] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 817.309752] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] nwinfo = self.network_api.allocate_for_instance( [ 817.309752] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 817.309752] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] created_port_ids = self._update_ports_for_instance( [ 817.309752] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 817.309752] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] with excutils.save_and_reraise_exception(): [ 817.309752] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 817.309752] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] self.force_reraise() [ 817.309752] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 817.310202] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] raise self.value [ 817.310202] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 817.310202] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] updated_port = self._update_port( [ 817.310202] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 817.310202] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] _ensure_no_port_binding_failure(port) [ 817.310202] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 817.310202] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] raise exception.PortBindingFailed(port_id=port['id']) [ 817.310202] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] nova.exception.PortBindingFailed: Binding failed for port 3a2c383f-766e-454a-bddf-a884d879da19, please check neutron logs for more information. [ 817.310202] env[61970]: ERROR nova.compute.manager [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] [ 817.310202] env[61970]: DEBUG nova.compute.utils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Binding failed for port 3a2c383f-766e-454a-bddf-a884d879da19, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 817.310872] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.674s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.314318] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Build of instance 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7 was re-scheduled: Binding failed for port 3a2c383f-766e-454a-bddf-a884d879da19, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 817.314739] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 817.314961] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquiring lock "refresh_cache-5ce0137d-a22e-4f22-9d3f-ed4812e67ec7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.315117] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquired lock "refresh_cache-5ce0137d-a22e-4f22-9d3f-ed4812e67ec7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.315274] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 817.356914] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355569, 'name': ReconfigVM_Task, 'duration_secs': 0.516725} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.357209] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Reconfigured VM instance instance-0000003a to attach disk [datastore2] a78b63a5-6bb8-4271-90d8-1e86fb29db4f/a78b63a5-6bb8-4271-90d8-1e86fb29db4f.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 817.357796] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9cc65708-99df-4895-8884-408df4b20f06 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.364131] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 817.364131] env[61970]: value = "task-1355570" [ 817.364131] env[61970]: _type = "Task" [ 817.364131] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.372044] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355570, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.452722] env[61970]: DEBUG nova.compute.manager [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 817.480910] env[61970]: DEBUG nova.virt.hardware [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 817.481184] env[61970]: DEBUG nova.virt.hardware [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 817.481340] env[61970]: DEBUG nova.virt.hardware [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 817.481517] env[61970]: DEBUG nova.virt.hardware [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 817.481657] env[61970]: DEBUG nova.virt.hardware [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 817.481801] env[61970]: DEBUG nova.virt.hardware [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 817.482026] env[61970]: DEBUG nova.virt.hardware [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 817.482202] env[61970]: DEBUG nova.virt.hardware [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 817.482372] env[61970]: DEBUG nova.virt.hardware [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 817.482949] env[61970]: DEBUG nova.virt.hardware [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 817.482949] env[61970]: DEBUG nova.virt.hardware [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 817.483881] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eac2aad-8695-4ad9-ac6b-af2ce3520463 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.491781] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-273ccb15-3f8b-4c58-ba72-96b2418f6f37 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.505790] env[61970]: ERROR nova.compute.manager [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9248c5b1-6dfe-4447-a7e4-805e208e2ce3, please check neutron logs for more information. [ 817.505790] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Traceback (most recent call last): [ 817.505790] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 817.505790] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] yield resources [ 817.505790] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 817.505790] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] self.driver.spawn(context, instance, image_meta, [ 817.505790] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 817.505790] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] self._vmops.spawn(context, instance, image_meta, injected_files, [ 817.505790] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 817.505790] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] vm_ref = self.build_virtual_machine(instance, [ 817.505790] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 817.506356] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] vif_infos = vmwarevif.get_vif_info(self._session, [ 817.506356] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 817.506356] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] for vif in network_info: [ 817.506356] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 817.506356] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] return self._sync_wrapper(fn, *args, **kwargs) [ 817.506356] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 817.506356] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] self.wait() [ 817.506356] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 817.506356] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] self[:] = self._gt.wait() [ 817.506356] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 817.506356] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] return self._exit_event.wait() [ 817.506356] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 817.506356] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] current.throw(*self._exc) [ 817.506697] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 817.506697] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] result = function(*args, **kwargs) [ 817.506697] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 817.506697] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] return func(*args, **kwargs) [ 817.506697] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 817.506697] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] raise e [ 817.506697] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 817.506697] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] nwinfo = self.network_api.allocate_for_instance( [ 817.506697] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 817.506697] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] created_port_ids = self._update_ports_for_instance( [ 817.506697] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 817.506697] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] with excutils.save_and_reraise_exception(): [ 817.506697] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 817.507010] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] self.force_reraise() [ 817.507010] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 817.507010] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] raise self.value [ 817.507010] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 817.507010] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] updated_port = self._update_port( [ 817.507010] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 817.507010] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] _ensure_no_port_binding_failure(port) [ 817.507010] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 817.507010] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] raise exception.PortBindingFailed(port_id=port['id']) [ 817.507010] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] nova.exception.PortBindingFailed: Binding failed for port 9248c5b1-6dfe-4447-a7e4-805e208e2ce3, please check neutron logs for more information. [ 817.507010] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] [ 817.507010] env[61970]: INFO nova.compute.manager [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Terminating instance [ 817.508461] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Acquiring lock "refresh_cache-27996960-d2d4-496c-b52d-5d7312aed040" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.576080] env[61970]: DEBUG nova.network.neutron [req-90e0ba58-e434-42e8-87c5-a618a82b60c2 req-ffdc6eff-3db4-4d0d-b634-ec0c15c0f04b service nova] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.650841] env[61970]: DEBUG nova.network.neutron [req-90e0ba58-e434-42e8-87c5-a618a82b60c2 req-ffdc6eff-3db4-4d0d-b634-ec0c15c0f04b service nova] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.660859] env[61970]: DEBUG oslo_concurrency.lockutils [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.838777] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.876961] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355570, 'name': Rename_Task, 'duration_secs': 0.133292} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.877262] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 817.877580] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c692d616-725b-42ab-85e1-5738f00ac593 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.884540] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 817.884540] env[61970]: value = "task-1355571" [ 817.884540] env[61970]: _type = "Task" [ 817.884540] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.891836] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355571, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.935089] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.117328] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93da4a6e-ecaa-464f-8bac-f6ba80b23f8c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.124274] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1892b42-b34a-4307-a527-dd474745f6c6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.153775] env[61970]: DEBUG oslo_concurrency.lockutils [req-90e0ba58-e434-42e8-87c5-a618a82b60c2 req-ffdc6eff-3db4-4d0d-b634-ec0c15c0f04b service nova] Releasing lock "refresh_cache-27996960-d2d4-496c-b52d-5d7312aed040" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.154387] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Acquired lock "refresh_cache-27996960-d2d4-496c-b52d-5d7312aed040" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.154608] env[61970]: DEBUG nova.network.neutron [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 818.156296] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae5b663a-a7ba-4dba-ac0a-d0ded3dde178 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.164413] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30925c16-fe24-4566-b299-3bfbd4432836 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.179242] env[61970]: DEBUG nova.compute.provider_tree [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.396234] env[61970]: DEBUG oslo_vmware.api [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355571, 'name': PowerOnVM_Task, 'duration_secs': 0.432364} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.399898] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 818.399898] env[61970]: INFO nova.compute.manager [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Took 5.30 seconds to spawn the instance on the hypervisor. [ 818.399898] env[61970]: DEBUG nova.compute.manager [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 818.399898] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-426a067f-2006-4510-9550-9e3e7bd6ead8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.439395] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Releasing lock "refresh_cache-5ce0137d-a22e-4f22-9d3f-ed4812e67ec7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.439395] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 818.439561] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 818.441034] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 818.456310] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.676221] env[61970]: DEBUG nova.network.neutron [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.682433] env[61970]: DEBUG nova.scheduler.client.report [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.810181] env[61970]: DEBUG nova.network.neutron [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.918051] env[61970]: INFO nova.compute.manager [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Took 37.63 seconds to build instance. [ 818.959525] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.147729] env[61970]: INFO nova.compute.manager [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Rebuilding instance [ 819.189751] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.879s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.191472] env[61970]: ERROR nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c2e6332a-9fc1-4dbb-aea3-77565735c463, please check neutron logs for more information. [ 819.191472] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Traceback (most recent call last): [ 819.191472] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 819.191472] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] self.driver.spawn(context, instance, image_meta, [ 819.191472] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 819.191472] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] self._vmops.spawn(context, instance, image_meta, injected_files, [ 819.191472] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 819.191472] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] vm_ref = self.build_virtual_machine(instance, [ 819.191472] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 819.191472] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] vif_infos = vmwarevif.get_vif_info(self._session, [ 819.191472] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 819.191939] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] for vif in network_info: [ 819.191939] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 819.191939] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] return self._sync_wrapper(fn, *args, **kwargs) [ 819.191939] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 819.191939] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] self.wait() [ 819.191939] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 819.191939] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] self[:] = self._gt.wait() [ 819.191939] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 819.191939] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] return self._exit_event.wait() [ 819.191939] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 819.191939] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] current.throw(*self._exc) [ 819.191939] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 819.191939] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] result = function(*args, **kwargs) [ 819.192425] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 819.192425] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] return func(*args, **kwargs) [ 819.192425] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 819.192425] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] raise e [ 819.192425] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 819.192425] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] nwinfo = self.network_api.allocate_for_instance( [ 819.192425] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 819.192425] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] created_port_ids = self._update_ports_for_instance( [ 819.192425] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 819.192425] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] with excutils.save_and_reraise_exception(): [ 819.192425] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 819.192425] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] self.force_reraise() [ 819.192425] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 819.192955] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] raise self.value [ 819.192955] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 819.192955] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] updated_port = self._update_port( [ 819.192955] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 819.192955] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] _ensure_no_port_binding_failure(port) [ 819.192955] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 819.192955] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] raise exception.PortBindingFailed(port_id=port['id']) [ 819.192955] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] nova.exception.PortBindingFailed: Binding failed for port c2e6332a-9fc1-4dbb-aea3-77565735c463, please check neutron logs for more information. [ 819.192955] env[61970]: ERROR nova.compute.manager [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] [ 819.192955] env[61970]: DEBUG nova.compute.utils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Binding failed for port c2e6332a-9fc1-4dbb-aea3-77565735c463, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 819.199947] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.814s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.207019] env[61970]: DEBUG nova.compute.manager [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 819.207019] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d03375-6893-448f-806e-117326360f57 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.222019] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Build of instance 3528d9b6-f3fb-43e9-be27-113d887a3414 was re-scheduled: Binding failed for port c2e6332a-9fc1-4dbb-aea3-77565735c463, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 819.222019] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 819.222203] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquiring lock "refresh_cache-3528d9b6-f3fb-43e9-be27-113d887a3414" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.222249] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquired lock "refresh_cache-3528d9b6-f3fb-43e9-be27-113d887a3414" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.222382] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 819.246995] env[61970]: DEBUG nova.compute.manager [req-6fed234b-f322-4447-b97a-ca37b09bf760 req-8680e931-543c-4eb8-b463-83dfd04ce2c0 service nova] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Received event network-vif-deleted-9248c5b1-6dfe-4447-a7e4-805e208e2ce3 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.312850] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Releasing lock "refresh_cache-27996960-d2d4-496c-b52d-5d7312aed040" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.313304] env[61970]: DEBUG nova.compute.manager [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 819.313494] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 819.314223] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ccb65780-0667-4a61-b1f7-3fea1c65465e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.323089] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429382e3-a3f6-4633-9f2a-0804581359d3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.345659] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 27996960-d2d4-496c-b52d-5d7312aed040 could not be found. [ 819.345899] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 819.346108] env[61970]: INFO nova.compute.manager [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Took 0.03 seconds to destroy the instance on the hypervisor. [ 819.346358] env[61970]: DEBUG oslo.service.loopingcall [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 819.346587] env[61970]: DEBUG nova.compute.manager [-] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 819.346680] env[61970]: DEBUG nova.network.neutron [-] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 819.362570] env[61970]: DEBUG nova.network.neutron [-] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 819.421420] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f615fbdb-6e29-4b72-9156-c400bfd9aeb7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "a78b63a5-6bb8-4271-90d8-1e86fb29db4f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.303s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.466918] env[61970]: INFO nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7] Took 1.03 seconds to deallocate network for instance. [ 819.719582] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 819.719869] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d294e43-a9ee-429c-b8ca-4c28e5e91d30 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.730973] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 819.730973] env[61970]: value = "task-1355572" [ 819.730973] env[61970]: _type = "Task" [ 819.730973] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.740130] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355572, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.754402] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 819.864963] env[61970]: DEBUG nova.network.neutron [-] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.905468] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.922684] env[61970]: DEBUG nova.compute.manager [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 820.068037] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-675e044a-b013-49da-960a-a3ddb8f7526a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.076231] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f251f2b-d06d-4530-b4fb-f72c228cbaf5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.112022] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1bb964f-5cc5-4044-889c-3604c9d417c0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.117493] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a1a04c-ffb6-449d-b334-b2accc289343 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.133776] env[61970]: DEBUG nova.compute.provider_tree [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.242731] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355572, 'name': PowerOffVM_Task, 'duration_secs': 0.125652} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.243486] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 820.243486] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 820.244077] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf3edbdb-d2db-421d-8915-93825b177fae {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.251451] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 820.251451] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6201ffae-ec22-43a7-a4b5-b4af683e1e96 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.277865] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 820.277865] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 820.278080] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Deleting the datastore file [datastore2] a78b63a5-6bb8-4271-90d8-1e86fb29db4f {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 820.278932] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-24bdd129-c606-4aea-9a9c-f412a64a5687 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.285407] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 820.285407] env[61970]: value = "task-1355574" [ 820.285407] env[61970]: _type = "Task" [ 820.285407] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.294243] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355574, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.367447] env[61970]: INFO nova.compute.manager [-] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Took 1.02 seconds to deallocate network for instance. [ 820.373377] env[61970]: DEBUG nova.compute.claims [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 820.373573] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.407876] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Releasing lock "refresh_cache-3528d9b6-f3fb-43e9-be27-113d887a3414" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.408134] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 820.408321] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 820.408489] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 820.428020] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.449818] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.496310] env[61970]: INFO nova.scheduler.client.report [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Deleted allocations for instance 5ce0137d-a22e-4f22-9d3f-ed4812e67ec7 [ 820.635175] env[61970]: DEBUG nova.scheduler.client.report [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 820.796568] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355574, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.109368} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.796833] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 820.797020] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 820.797198] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 820.929818] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.009545] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Lock "5ce0137d-a22e-4f22-9d3f-ed4812e67ec7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 158.626s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.141931] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.941s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.142052] env[61970]: ERROR nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7434460f-bcc3-4312-8a89-15154d936f92, please check neutron logs for more information. [ 821.142052] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Traceback (most recent call last): [ 821.142052] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 821.142052] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] self.driver.spawn(context, instance, image_meta, [ 821.142052] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 821.142052] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 821.142052] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 821.142052] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] vm_ref = self.build_virtual_machine(instance, [ 821.142052] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 821.142052] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] vif_infos = vmwarevif.get_vif_info(self._session, [ 821.142052] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 821.142380] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] for vif in network_info: [ 821.142380] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 821.142380] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] return self._sync_wrapper(fn, *args, **kwargs) [ 821.142380] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 821.142380] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] self.wait() [ 821.142380] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 821.142380] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] self[:] = self._gt.wait() [ 821.142380] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 821.142380] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] return self._exit_event.wait() [ 821.142380] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 821.142380] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] current.throw(*self._exc) [ 821.142380] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 821.142380] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] result = function(*args, **kwargs) [ 821.142727] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 821.142727] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] return func(*args, **kwargs) [ 821.142727] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 821.142727] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] raise e [ 821.142727] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 821.142727] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] nwinfo = self.network_api.allocate_for_instance( [ 821.142727] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 821.142727] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] created_port_ids = self._update_ports_for_instance( [ 821.142727] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 821.142727] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] with excutils.save_and_reraise_exception(): [ 821.142727] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 821.142727] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] self.force_reraise() [ 821.142727] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 821.143100] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] raise self.value [ 821.143100] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 821.143100] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] updated_port = self._update_port( [ 821.143100] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 821.143100] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] _ensure_no_port_binding_failure(port) [ 821.143100] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 821.143100] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] raise exception.PortBindingFailed(port_id=port['id']) [ 821.143100] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] nova.exception.PortBindingFailed: Binding failed for port 7434460f-bcc3-4312-8a89-15154d936f92, please check neutron logs for more information. [ 821.143100] env[61970]: ERROR nova.compute.manager [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] [ 821.143100] env[61970]: DEBUG nova.compute.utils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Binding failed for port 7434460f-bcc3-4312-8a89-15154d936f92, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 821.144032] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.441s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.147750] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Build of instance 07eeb310-312b-4f32-bd5f-7b254cff0fe1 was re-scheduled: Binding failed for port 7434460f-bcc3-4312-8a89-15154d936f92, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 821.148308] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 821.148546] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquiring lock "refresh_cache-07eeb310-312b-4f32-bd5f-7b254cff0fe1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.148696] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Acquired lock "refresh_cache-07eeb310-312b-4f32-bd5f-7b254cff0fe1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.148872] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 821.433629] env[61970]: INFO nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 3528d9b6-f3fb-43e9-be27-113d887a3414] Took 1.02 seconds to deallocate network for instance. [ 821.512057] env[61970]: DEBUG nova.compute.manager [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 821.675523] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 821.784904] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.840018] env[61970]: DEBUG nova.virt.hardware [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 821.840018] env[61970]: DEBUG nova.virt.hardware [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 821.840018] env[61970]: DEBUG nova.virt.hardware [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.840018] env[61970]: DEBUG nova.virt.hardware [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 821.840290] env[61970]: DEBUG nova.virt.hardware [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.840290] env[61970]: DEBUG nova.virt.hardware [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 821.840880] env[61970]: DEBUG nova.virt.hardware [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 821.841273] env[61970]: DEBUG nova.virt.hardware [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 821.843045] env[61970]: DEBUG nova.virt.hardware [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 821.843045] env[61970]: DEBUG nova.virt.hardware [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 821.843045] env[61970]: DEBUG nova.virt.hardware [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 821.843711] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd0080d-570d-41b8-8765-4f5b36738cef {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.853969] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08cf03f2-d7ce-48a2-ac19-24ad8909865d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.870165] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Instance VIF info [] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 821.875783] env[61970]: DEBUG oslo.service.loopingcall [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.876090] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 821.876338] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-27effd5a-883a-4bac-a66f-6ba274c09d99 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.895186] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 821.895186] env[61970]: value = "task-1355575" [ 821.895186] env[61970]: _type = "Task" [ 821.895186] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.903320] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355575, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.973174] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff33bb98-534c-41be-a7d8-f54369258f04 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.982957] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03bd4d5-a072-4242-b5a8-1f311cefffa2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.019117] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa0e9d8-b45d-44d8-bf2e-05f98537df77 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.029530] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-419cd98c-d5c0-4e2a-8110-b7de7fd26e40 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.043606] env[61970]: DEBUG nova.compute.provider_tree [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.045352] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.287521] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Releasing lock "refresh_cache-07eeb310-312b-4f32-bd5f-7b254cff0fe1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.287748] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 822.287931] env[61970]: DEBUG nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 822.288128] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 822.304964] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.406065] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355575, 'name': CreateVM_Task, 'duration_secs': 0.299806} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.406231] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 822.407270] env[61970]: DEBUG oslo_vmware.service [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aac00d19-cec1-4b1a-9d7b-c8bed3cb4311 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.413040] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.413482] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.413586] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 822.413834] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47a3ea8e-f40c-4552-bdc9-e36b0dd24dfe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.419832] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 822.419832] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ac3218-0856-39fa-65ea-e8c867998caf" [ 822.419832] env[61970]: _type = "Task" [ 822.419832] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.427074] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ac3218-0856-39fa-65ea-e8c867998caf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.466969] env[61970]: INFO nova.scheduler.client.report [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Deleted allocations for instance 3528d9b6-f3fb-43e9-be27-113d887a3414 [ 822.549020] env[61970]: DEBUG nova.scheduler.client.report [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 822.807928] env[61970]: DEBUG nova.network.neutron [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.931869] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.932157] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 822.932384] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.932536] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.932713] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 822.932971] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cdbfe1dc-40cc-4ef9-9fd3-06454dfa8bec {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.941045] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 822.941151] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 822.941835] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104ef3f6-a9a1-41b1-825a-4a69739b17ae {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.947806] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-118bd715-4f33-4ee3-b1a5-5c2156b63888 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.952479] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 822.952479] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b0513c-4075-97c2-b23f-509dd85145d3" [ 822.952479] env[61970]: _type = "Task" [ 822.952479] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.959803] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b0513c-4075-97c2-b23f-509dd85145d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.977694] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Lock "3528d9b6-f3fb-43e9-be27-113d887a3414" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 160.569s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.053144] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.909s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.053798] env[61970]: ERROR nova.compute.manager [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0ce24b8d-7846-42d7-817e-8118b9bff9ee, please check neutron logs for more information. [ 823.053798] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Traceback (most recent call last): [ 823.053798] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 823.053798] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] self.driver.spawn(context, instance, image_meta, [ 823.053798] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 823.053798] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] self._vmops.spawn(context, instance, image_meta, injected_files, [ 823.053798] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 823.053798] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] vm_ref = self.build_virtual_machine(instance, [ 823.053798] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 823.053798] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] vif_infos = vmwarevif.get_vif_info(self._session, [ 823.053798] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 823.054258] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] for vif in network_info: [ 823.054258] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 823.054258] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] return self._sync_wrapper(fn, *args, **kwargs) [ 823.054258] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 823.054258] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] self.wait() [ 823.054258] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 823.054258] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] self[:] = self._gt.wait() [ 823.054258] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 823.054258] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] return self._exit_event.wait() [ 823.054258] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 823.054258] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] current.throw(*self._exc) [ 823.054258] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 823.054258] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] result = function(*args, **kwargs) [ 823.054601] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 823.054601] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] return func(*args, **kwargs) [ 823.054601] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 823.054601] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] raise e [ 823.054601] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 823.054601] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] nwinfo = self.network_api.allocate_for_instance( [ 823.054601] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 823.054601] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] created_port_ids = self._update_ports_for_instance( [ 823.054601] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 823.054601] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] with excutils.save_and_reraise_exception(): [ 823.054601] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 823.054601] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] self.force_reraise() [ 823.054601] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 823.054940] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] raise self.value [ 823.054940] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 823.054940] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] updated_port = self._update_port( [ 823.054940] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 823.054940] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] _ensure_no_port_binding_failure(port) [ 823.054940] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 823.054940] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] raise exception.PortBindingFailed(port_id=port['id']) [ 823.054940] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] nova.exception.PortBindingFailed: Binding failed for port 0ce24b8d-7846-42d7-817e-8118b9bff9ee, please check neutron logs for more information. [ 823.054940] env[61970]: ERROR nova.compute.manager [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] [ 823.054940] env[61970]: DEBUG nova.compute.utils [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Binding failed for port 0ce24b8d-7846-42d7-817e-8118b9bff9ee, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 823.056846] env[61970]: DEBUG nova.compute.manager [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Build of instance a8ecafbe-f1e1-49f1-945d-ef8b6d15b709 was re-scheduled: Binding failed for port 0ce24b8d-7846-42d7-817e-8118b9bff9ee, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 823.056846] env[61970]: DEBUG nova.compute.manager [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 823.057199] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "refresh_cache-a8ecafbe-f1e1-49f1-945d-ef8b6d15b709" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.057199] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "refresh_cache-a8ecafbe-f1e1-49f1-945d-ef8b6d15b709" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.057275] env[61970]: DEBUG nova.network.neutron [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 823.060607] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.864s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.060607] env[61970]: INFO nova.compute.claims [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.311887] env[61970]: INFO nova.compute.manager [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] [instance: 07eeb310-312b-4f32-bd5f-7b254cff0fe1] Took 1.02 seconds to deallocate network for instance. [ 823.463184] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Preparing fetch location {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 823.463454] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Creating directory with path [datastore1] vmware_temp/a1f72f3e-d37c-4e6b-9809-47310873dcdf/7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 823.463698] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f1fa6d77-f7a6-4728-afc8-40f0c762c844 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.480425] env[61970]: DEBUG nova.compute.manager [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 823.484604] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Created directory with path [datastore1] vmware_temp/a1f72f3e-d37c-4e6b-9809-47310873dcdf/7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 823.484799] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Fetch image to [datastore1] vmware_temp/a1f72f3e-d37c-4e6b-9809-47310873dcdf/7ff5af44-3f73-4c84-82e5-a1543db26d03/tmp-sparse.vmdk {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 823.484969] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Downloading image file data 7ff5af44-3f73-4c84-82e5-a1543db26d03 to [datastore1] vmware_temp/a1f72f3e-d37c-4e6b-9809-47310873dcdf/7ff5af44-3f73-4c84-82e5-a1543db26d03/tmp-sparse.vmdk on the data store datastore1 {{(pid=61970) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 823.485930] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e811193-38d8-4d12-ace7-eb6d0ab69cb7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.492686] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43977e41-7546-4a33-abba-aa0932a2fcbc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.502587] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea39208-8705-4c59-bd5e-b64056deb34c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.534948] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60de3cb0-cb01-4dc9-b8d3-ce0776c476fc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.540787] env[61970]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-7683f76f-28fb-4d58-9654-334985559579 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.560221] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Downloading image file data 7ff5af44-3f73-4c84-82e5-a1543db26d03 to the data store datastore1 {{(pid=61970) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 823.586434] env[61970]: DEBUG nova.network.neutron [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.612028] env[61970]: DEBUG oslo_vmware.rw_handles [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/a1f72f3e-d37c-4e6b-9809-47310873dcdf/7ff5af44-3f73-4c84-82e5-a1543db26d03/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61970) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 823.709531] env[61970]: DEBUG nova.network.neutron [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.005509] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.213530] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "refresh_cache-a8ecafbe-f1e1-49f1-945d-ef8b6d15b709" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.213803] env[61970]: DEBUG nova.compute.manager [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 824.214061] env[61970]: DEBUG nova.compute.manager [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 824.214291] env[61970]: DEBUG nova.network.neutron [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 824.232790] env[61970]: DEBUG nova.network.neutron [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.282181] env[61970]: DEBUG oslo_vmware.rw_handles [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Completed reading data from the image iterator. {{(pid=61970) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 824.282406] env[61970]: DEBUG oslo_vmware.rw_handles [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/a1f72f3e-d37c-4e6b-9809-47310873dcdf/7ff5af44-3f73-4c84-82e5-a1543db26d03/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61970) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 824.314410] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-554fd619-3bf1-4dd0-9af9-9e9e78a28104 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.323906] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984822e5-fe79-4012-822c-1c95b1c3cdd3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.354068] env[61970]: INFO nova.scheduler.client.report [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Deleted allocations for instance 07eeb310-312b-4f32-bd5f-7b254cff0fe1 [ 824.359866] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f177d142-c6a6-4d38-88a8-0d4a872bb758 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.369050] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e147a823-cde0-4791-b6b7-36e04de096da {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.385375] env[61970]: DEBUG nova.compute.provider_tree [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.422082] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Downloaded image file data 7ff5af44-3f73-4c84-82e5-a1543db26d03 to vmware_temp/a1f72f3e-d37c-4e6b-9809-47310873dcdf/7ff5af44-3f73-4c84-82e5-a1543db26d03/tmp-sparse.vmdk on the data store datastore1 {{(pid=61970) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 824.423384] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Caching image {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 824.423637] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Copying Virtual Disk [datastore1] vmware_temp/a1f72f3e-d37c-4e6b-9809-47310873dcdf/7ff5af44-3f73-4c84-82e5-a1543db26d03/tmp-sparse.vmdk to [datastore1] vmware_temp/a1f72f3e-d37c-4e6b-9809-47310873dcdf/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 824.423896] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6656f9b2-265f-4dcc-ab35-b49b2d28fb91 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.431564] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 824.431564] env[61970]: value = "task-1355576" [ 824.431564] env[61970]: _type = "Task" [ 824.431564] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.439218] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355576, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.735957] env[61970]: DEBUG nova.network.neutron [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.865602] env[61970]: DEBUG oslo_concurrency.lockutils [None req-930fd6cd-c0d1-4252-8560-232a6bc65de1 tempest-ListServersNegativeTestJSON-1249336202 tempest-ListServersNegativeTestJSON-1249336202-project-member] Lock "07eeb310-312b-4f32-bd5f-7b254cff0fe1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 162.423s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.889493] env[61970]: DEBUG nova.scheduler.client.report [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 824.942598] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355576, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.238672] env[61970]: INFO nova.compute.manager [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: a8ecafbe-f1e1-49f1-945d-ef8b6d15b709] Took 1.02 seconds to deallocate network for instance. [ 825.368196] env[61970]: DEBUG nova.compute.manager [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 825.395176] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.337s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.395841] env[61970]: DEBUG nova.compute.manager [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 825.398397] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.929s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.400016] env[61970]: INFO nova.compute.claims [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 825.443324] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355576, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.664659} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.443706] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Copied Virtual Disk [datastore1] vmware_temp/a1f72f3e-d37c-4e6b-9809-47310873dcdf/7ff5af44-3f73-4c84-82e5-a1543db26d03/tmp-sparse.vmdk to [datastore1] vmware_temp/a1f72f3e-d37c-4e6b-9809-47310873dcdf/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 825.443824] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Deleting the datastore file [datastore1] vmware_temp/a1f72f3e-d37c-4e6b-9809-47310873dcdf/7ff5af44-3f73-4c84-82e5-a1543db26d03/tmp-sparse.vmdk {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 825.444070] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b653a73f-2fc0-4a8c-8175-f23ce1a7595e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.450487] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 825.450487] env[61970]: value = "task-1355577" [ 825.450487] env[61970]: _type = "Task" [ 825.450487] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.457654] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355577, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.889356] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.905946] env[61970]: DEBUG nova.compute.utils [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 825.910797] env[61970]: DEBUG nova.compute.manager [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 825.910797] env[61970]: DEBUG nova.network.neutron [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 825.967664] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355577, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.022517} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.969209] env[61970]: DEBUG nova.policy [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8622ff7392ba4fadaf655457f42965d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd23e0fee8480406cbdc5011db462a5b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 825.970696] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 825.970903] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Moving file from [datastore1] vmware_temp/a1f72f3e-d37c-4e6b-9809-47310873dcdf/7ff5af44-3f73-4c84-82e5-a1543db26d03 to [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03. {{(pid=61970) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 825.971184] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-2448f69a-1e6f-4b31-a930-aba90f6409d0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.978639] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 825.978639] env[61970]: value = "task-1355578" [ 825.978639] env[61970]: _type = "Task" [ 825.978639] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.989554] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355578, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.258656] env[61970]: DEBUG nova.network.neutron [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Successfully created port: 1187b71f-0e90-465d-b15a-d9eae922f9de {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 826.281763] env[61970]: INFO nova.scheduler.client.report [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Deleted allocations for instance a8ecafbe-f1e1-49f1-945d-ef8b6d15b709 [ 826.409175] env[61970]: DEBUG nova.compute.manager [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 826.492986] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355578, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.023908} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.499671] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] File moved {{(pid=61970) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 826.499893] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Cleaning up location [datastore1] vmware_temp/a1f72f3e-d37c-4e6b-9809-47310873dcdf {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 826.500066] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Deleting the datastore file [datastore1] vmware_temp/a1f72f3e-d37c-4e6b-9809-47310873dcdf {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 826.500603] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db0343be-6b5f-40b3-9dfe-969aed8e5aef {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.508582] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 826.508582] env[61970]: value = "task-1355579" [ 826.508582] env[61970]: _type = "Task" [ 826.508582] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.516939] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355579, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.698104] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4994b4-e052-40e7-b66f-566f13d941f6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.704982] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45514aae-4740-4bbc-8e26-6b4b784eca34 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.738115] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c43984e8-d0ad-4767-8869-b9fd739be53a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.745828] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c9f3ae0-f627-45fd-8348-2f4fee5c9e7b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.760280] env[61970]: DEBUG nova.compute.provider_tree [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.790042] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f8379c18-cab2-4ffe-b9b5-51567c45ed4c tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "a8ecafbe-f1e1-49f1-945d-ef8b6d15b709" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 161.203s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.020106] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355579, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.028193} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.020416] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 827.021154] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e62fbd5-67ba-428e-a6a6-66166ecd8393 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.030704] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 827.030704] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]524ff800-e84b-7098-5518-c0a79ae6eb42" [ 827.030704] env[61970]: _type = "Task" [ 827.030704] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.042781] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]524ff800-e84b-7098-5518-c0a79ae6eb42, 'name': SearchDatastore_Task, 'duration_secs': 0.010087} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.043061] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.043437] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] a78b63a5-6bb8-4271-90d8-1e86fb29db4f/a78b63a5-6bb8-4271-90d8-1e86fb29db4f.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 827.043715] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b95f65e8-4699-4235-a778-2fe961315542 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.049802] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 827.049802] env[61970]: value = "task-1355580" [ 827.049802] env[61970]: _type = "Task" [ 827.049802] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.057620] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355580, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.087080] env[61970]: DEBUG nova.compute.manager [req-5f833205-2fd9-4f70-b1e1-0f90549d0f90 req-613e6188-4e72-497a-bfb4-dc18277de420 service nova] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Received event network-changed-1187b71f-0e90-465d-b15a-d9eae922f9de {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.087333] env[61970]: DEBUG nova.compute.manager [req-5f833205-2fd9-4f70-b1e1-0f90549d0f90 req-613e6188-4e72-497a-bfb4-dc18277de420 service nova] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Refreshing instance network info cache due to event network-changed-1187b71f-0e90-465d-b15a-d9eae922f9de. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 827.087615] env[61970]: DEBUG oslo_concurrency.lockutils [req-5f833205-2fd9-4f70-b1e1-0f90549d0f90 req-613e6188-4e72-497a-bfb4-dc18277de420 service nova] Acquiring lock "refresh_cache-c5c253b6-ebab-4fea-bcc4-f3bb3960070b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.087823] env[61970]: DEBUG oslo_concurrency.lockutils [req-5f833205-2fd9-4f70-b1e1-0f90549d0f90 req-613e6188-4e72-497a-bfb4-dc18277de420 service nova] Acquired lock "refresh_cache-c5c253b6-ebab-4fea-bcc4-f3bb3960070b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.088618] env[61970]: DEBUG nova.network.neutron [req-5f833205-2fd9-4f70-b1e1-0f90549d0f90 req-613e6188-4e72-497a-bfb4-dc18277de420 service nova] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Refreshing network info cache for port 1187b71f-0e90-465d-b15a-d9eae922f9de {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 827.266249] env[61970]: DEBUG nova.scheduler.client.report [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 827.297549] env[61970]: DEBUG nova.compute.manager [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 827.325390] env[61970]: ERROR nova.compute.manager [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1187b71f-0e90-465d-b15a-d9eae922f9de, please check neutron logs for more information. [ 827.325390] env[61970]: ERROR nova.compute.manager Traceback (most recent call last): [ 827.325390] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 827.325390] env[61970]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 827.325390] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 827.325390] env[61970]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 827.325390] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 827.325390] env[61970]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 827.325390] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 827.325390] env[61970]: ERROR nova.compute.manager self.force_reraise() [ 827.325390] env[61970]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 827.325390] env[61970]: ERROR nova.compute.manager raise self.value [ 827.325390] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 827.325390] env[61970]: ERROR nova.compute.manager updated_port = self._update_port( [ 827.325390] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 827.325390] env[61970]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 827.326024] env[61970]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 827.326024] env[61970]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 827.326024] env[61970]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1187b71f-0e90-465d-b15a-d9eae922f9de, please check neutron logs for more information. [ 827.326024] env[61970]: ERROR nova.compute.manager [ 827.326024] env[61970]: Traceback (most recent call last): [ 827.326024] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 827.326024] env[61970]: listener.cb(fileno) [ 827.326024] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 827.326024] env[61970]: result = function(*args, **kwargs) [ 827.326024] env[61970]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 827.326024] env[61970]: return func(*args, **kwargs) [ 827.326024] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 827.326024] env[61970]: raise e [ 827.326024] env[61970]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 827.326024] env[61970]: nwinfo = self.network_api.allocate_for_instance( [ 827.326024] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 827.326024] env[61970]: created_port_ids = self._update_ports_for_instance( [ 827.326024] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 827.326024] env[61970]: with excutils.save_and_reraise_exception(): [ 827.326024] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 827.326024] env[61970]: self.force_reraise() [ 827.326024] env[61970]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 827.326024] env[61970]: raise self.value [ 827.326024] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 827.326024] env[61970]: updated_port = self._update_port( [ 827.326024] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 827.326024] env[61970]: _ensure_no_port_binding_failure(port) [ 827.326024] env[61970]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 827.326024] env[61970]: raise exception.PortBindingFailed(port_id=port['id']) [ 827.326680] env[61970]: nova.exception.PortBindingFailed: Binding failed for port 1187b71f-0e90-465d-b15a-d9eae922f9de, please check neutron logs for more information. [ 827.326680] env[61970]: Removing descriptor: 17 [ 827.423433] env[61970]: DEBUG nova.compute.manager [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 827.450749] env[61970]: DEBUG nova.virt.hardware [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 827.451060] env[61970]: DEBUG nova.virt.hardware [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 827.451240] env[61970]: DEBUG nova.virt.hardware [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 827.451426] env[61970]: DEBUG nova.virt.hardware [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 827.451585] env[61970]: DEBUG nova.virt.hardware [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 827.451733] env[61970]: DEBUG nova.virt.hardware [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 827.451939] env[61970]: DEBUG nova.virt.hardware [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 827.452134] env[61970]: DEBUG nova.virt.hardware [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 827.452272] env[61970]: DEBUG nova.virt.hardware [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 827.452458] env[61970]: DEBUG nova.virt.hardware [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 827.452779] env[61970]: DEBUG nova.virt.hardware [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 827.453705] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f85b3451-c690-4b74-9372-f8bfea73a27f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.464851] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d648a5d-3d4a-457c-bc13-239b1e9cfcf1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.480824] env[61970]: ERROR nova.compute.manager [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1187b71f-0e90-465d-b15a-d9eae922f9de, please check neutron logs for more information. [ 827.480824] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Traceback (most recent call last): [ 827.480824] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 827.480824] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] yield resources [ 827.480824] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 827.480824] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] self.driver.spawn(context, instance, image_meta, [ 827.480824] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 827.480824] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 827.480824] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 827.480824] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] vm_ref = self.build_virtual_machine(instance, [ 827.480824] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 827.481197] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] vif_infos = vmwarevif.get_vif_info(self._session, [ 827.481197] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 827.481197] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] for vif in network_info: [ 827.481197] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 827.481197] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] return self._sync_wrapper(fn, *args, **kwargs) [ 827.481197] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 827.481197] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] self.wait() [ 827.481197] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 827.481197] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] self[:] = self._gt.wait() [ 827.481197] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 827.481197] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] return self._exit_event.wait() [ 827.481197] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 827.481197] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] current.throw(*self._exc) [ 827.481513] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 827.481513] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] result = function(*args, **kwargs) [ 827.481513] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 827.481513] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] return func(*args, **kwargs) [ 827.481513] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 827.481513] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] raise e [ 827.481513] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 827.481513] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] nwinfo = self.network_api.allocate_for_instance( [ 827.481513] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 827.481513] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] created_port_ids = self._update_ports_for_instance( [ 827.481513] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 827.481513] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] with excutils.save_and_reraise_exception(): [ 827.481513] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 827.481796] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] self.force_reraise() [ 827.481796] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 827.481796] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] raise self.value [ 827.481796] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 827.481796] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] updated_port = self._update_port( [ 827.481796] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 827.481796] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] _ensure_no_port_binding_failure(port) [ 827.481796] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 827.481796] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] raise exception.PortBindingFailed(port_id=port['id']) [ 827.481796] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] nova.exception.PortBindingFailed: Binding failed for port 1187b71f-0e90-465d-b15a-d9eae922f9de, please check neutron logs for more information. [ 827.481796] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] [ 827.481796] env[61970]: INFO nova.compute.manager [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Terminating instance [ 827.483723] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Acquiring lock "refresh_cache-c5c253b6-ebab-4fea-bcc4-f3bb3960070b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.559533] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355580, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.607443] env[61970]: DEBUG nova.network.neutron [req-5f833205-2fd9-4f70-b1e1-0f90549d0f90 req-613e6188-4e72-497a-bfb4-dc18277de420 service nova] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 827.690285] env[61970]: DEBUG nova.network.neutron [req-5f833205-2fd9-4f70-b1e1-0f90549d0f90 req-613e6188-4e72-497a-bfb4-dc18277de420 service nova] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.771629] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.772207] env[61970]: DEBUG nova.compute.manager [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 827.775074] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.529s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.776455] env[61970]: INFO nova.compute.claims [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 827.821797] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.063847] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355580, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.678232} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.064165] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] a78b63a5-6bb8-4271-90d8-1e86fb29db4f/a78b63a5-6bb8-4271-90d8-1e86fb29db4f.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 828.064363] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.064619] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fee8fbe5-8e97-4fba-859b-332da30a789e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.071068] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 828.071068] env[61970]: value = "task-1355581" [ 828.071068] env[61970]: _type = "Task" [ 828.071068] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.082058] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355581, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.193449] env[61970]: DEBUG oslo_concurrency.lockutils [req-5f833205-2fd9-4f70-b1e1-0f90549d0f90 req-613e6188-4e72-497a-bfb4-dc18277de420 service nova] Releasing lock "refresh_cache-c5c253b6-ebab-4fea-bcc4-f3bb3960070b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.193988] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Acquired lock "refresh_cache-c5c253b6-ebab-4fea-bcc4-f3bb3960070b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.194283] env[61970]: DEBUG nova.network.neutron [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 828.280719] env[61970]: DEBUG nova.compute.utils [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 828.283982] env[61970]: DEBUG nova.compute.manager [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 828.284137] env[61970]: DEBUG nova.network.neutron [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 828.323559] env[61970]: DEBUG nova.policy [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98de02a5c4e14e9889be2fa1637b68d7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e91895bd10c74f8d9a101274d87cdc71', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 828.552542] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "75762ba3-f129-424e-a7cb-962785dfab2a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.554921] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "75762ba3-f129-424e-a7cb-962785dfab2a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.583062] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355581, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062892} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.583376] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 828.584155] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0a16c1-c254-4794-aec0-51b236381c7d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.604040] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] a78b63a5-6bb8-4271-90d8-1e86fb29db4f/a78b63a5-6bb8-4271-90d8-1e86fb29db4f.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 828.605034] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e81ae4dc-79ed-4ce7-9fd0-329ac559c727 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.626140] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 828.626140] env[61970]: value = "task-1355582" [ 828.626140] env[61970]: _type = "Task" [ 828.626140] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.636681] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355582, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.727746] env[61970]: DEBUG nova.network.neutron [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 828.731540] env[61970]: DEBUG nova.network.neutron [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Successfully created port: 1080e2f3-c9a3-48bc-bd65-b5f231bacdc7 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 828.784557] env[61970]: DEBUG nova.compute.manager [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 829.057158] env[61970]: DEBUG nova.network.neutron [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.087767] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18b4ef28-ba4c-4379-84c2-b09fd782c4c8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.096582] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb2ee12-dc35-4aea-90ac-3a5fdf1ed827 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.138435] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01a193f5-4ce5-42da-88c4-a29d07d67ce8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.142496] env[61970]: DEBUG nova.compute.manager [req-2e9b07ce-bdb4-474a-900c-19628af451dd req-3b407605-a44e-4876-ad70-ea7dbd5c81c1 service nova] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Received event network-vif-deleted-1187b71f-0e90-465d-b15a-d9eae922f9de {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.148476] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355582, 'name': ReconfigVM_Task, 'duration_secs': 0.304182} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.153023] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Reconfigured VM instance instance-0000003a to attach disk [datastore1] a78b63a5-6bb8-4271-90d8-1e86fb29db4f/a78b63a5-6bb8-4271-90d8-1e86fb29db4f.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.153023] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a645ee04-9110-4fb2-a1ac-2c8554a71adf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.154567] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb211559-e749-4d59-8e1b-c45db0d35eb2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.173849] env[61970]: DEBUG nova.compute.provider_tree [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.176514] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 829.176514] env[61970]: value = "task-1355583" [ 829.176514] env[61970]: _type = "Task" [ 829.176514] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.184758] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355583, 'name': Rename_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.560208] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Releasing lock "refresh_cache-c5c253b6-ebab-4fea-bcc4-f3bb3960070b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.560208] env[61970]: DEBUG nova.compute.manager [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 829.560711] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 829.561181] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bc15f1d2-5a27-43a0-b289-fcfdf415fd75 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.570928] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a1f90b-6b27-474d-9c9c-68b4d2df2840 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.593229] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c5c253b6-ebab-4fea-bcc4-f3bb3960070b could not be found. [ 829.593486] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 829.593735] env[61970]: INFO nova.compute.manager [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 829.594079] env[61970]: DEBUG oslo.service.loopingcall [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.594410] env[61970]: DEBUG nova.compute.manager [-] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 829.594562] env[61970]: DEBUG nova.network.neutron [-] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 829.612602] env[61970]: DEBUG nova.network.neutron [-] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 829.678347] env[61970]: DEBUG nova.scheduler.client.report [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 829.691368] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355583, 'name': Rename_Task, 'duration_secs': 0.155642} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.691645] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 829.691965] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c589c10d-cf8d-4f9c-97f1-4a1e28df7bc3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.698605] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 829.698605] env[61970]: value = "task-1355584" [ 829.698605] env[61970]: _type = "Task" [ 829.698605] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.708010] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355584, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.796959] env[61970]: DEBUG nova.compute.manager [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 829.827967] env[61970]: DEBUG nova.virt.hardware [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 829.828660] env[61970]: DEBUG nova.virt.hardware [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 829.828660] env[61970]: DEBUG nova.virt.hardware [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 829.828660] env[61970]: DEBUG nova.virt.hardware [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 829.828901] env[61970]: DEBUG nova.virt.hardware [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 829.829926] env[61970]: DEBUG nova.virt.hardware [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 829.829926] env[61970]: DEBUG nova.virt.hardware [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 829.829926] env[61970]: DEBUG nova.virt.hardware [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 829.829926] env[61970]: DEBUG nova.virt.hardware [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 829.829926] env[61970]: DEBUG nova.virt.hardware [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 829.831109] env[61970]: DEBUG nova.virt.hardware [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 829.832463] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac0341c-addf-4a39-b515-745d9eb37b92 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.840749] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd0d4c0-2776-4460-bd10-a6a19e635544 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.115863] env[61970]: DEBUG nova.network.neutron [-] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.187503] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.188235] env[61970]: DEBUG nova.compute.manager [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 830.193834] env[61970]: DEBUG oslo_concurrency.lockutils [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.533s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.195443] env[61970]: INFO nova.compute.claims [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 830.209094] env[61970]: DEBUG oslo_vmware.api [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355584, 'name': PowerOnVM_Task, 'duration_secs': 0.429881} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.209427] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 830.211467] env[61970]: DEBUG nova.compute.manager [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 830.214523] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62f02e3c-da8f-4cf1-b089-82a9bc50274d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.623074] env[61970]: INFO nova.compute.manager [-] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Took 1.03 seconds to deallocate network for instance. [ 830.629764] env[61970]: DEBUG nova.compute.claims [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Aborting claim: {{(pid=61970) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 830.629839] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.702423] env[61970]: DEBUG nova.compute.utils [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 830.703806] env[61970]: DEBUG nova.compute.manager [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 830.703970] env[61970]: DEBUG nova.network.neutron [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 830.734383] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.760678] env[61970]: DEBUG nova.network.neutron [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Successfully updated port: 1080e2f3-c9a3-48bc-bd65-b5f231bacdc7 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 830.786807] env[61970]: DEBUG nova.policy [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98de02a5c4e14e9889be2fa1637b68d7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e91895bd10c74f8d9a101274d87cdc71', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 831.160363] env[61970]: DEBUG nova.network.neutron [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Successfully created port: c2860647-20d9-411e-9d79-e26336519408 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 831.167742] env[61970]: DEBUG nova.compute.manager [req-0b28bae8-107a-4541-a9e6-1a73d0d63f29 req-a404d142-ee30-4160-adf3-4595a05dba1e service nova] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Received event network-vif-plugged-1080e2f3-c9a3-48bc-bd65-b5f231bacdc7 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.168115] env[61970]: DEBUG oslo_concurrency.lockutils [req-0b28bae8-107a-4541-a9e6-1a73d0d63f29 req-a404d142-ee30-4160-adf3-4595a05dba1e service nova] Acquiring lock "5d4d97e1-5812-4884-a1ea-6eb0cee591c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.168922] env[61970]: DEBUG oslo_concurrency.lockutils [req-0b28bae8-107a-4541-a9e6-1a73d0d63f29 req-a404d142-ee30-4160-adf3-4595a05dba1e service nova] Lock "5d4d97e1-5812-4884-a1ea-6eb0cee591c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.168980] env[61970]: DEBUG oslo_concurrency.lockutils [req-0b28bae8-107a-4541-a9e6-1a73d0d63f29 req-a404d142-ee30-4160-adf3-4595a05dba1e service nova] Lock "5d4d97e1-5812-4884-a1ea-6eb0cee591c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.169257] env[61970]: DEBUG nova.compute.manager [req-0b28bae8-107a-4541-a9e6-1a73d0d63f29 req-a404d142-ee30-4160-adf3-4595a05dba1e service nova] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] No waiting events found dispatching network-vif-plugged-1080e2f3-c9a3-48bc-bd65-b5f231bacdc7 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 831.169480] env[61970]: WARNING nova.compute.manager [req-0b28bae8-107a-4541-a9e6-1a73d0d63f29 req-a404d142-ee30-4160-adf3-4595a05dba1e service nova] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Received unexpected event network-vif-plugged-1080e2f3-c9a3-48bc-bd65-b5f231bacdc7 for instance with vm_state building and task_state spawning. [ 831.169699] env[61970]: DEBUG nova.compute.manager [req-0b28bae8-107a-4541-a9e6-1a73d0d63f29 req-a404d142-ee30-4160-adf3-4595a05dba1e service nova] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Received event network-changed-1080e2f3-c9a3-48bc-bd65-b5f231bacdc7 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.169980] env[61970]: DEBUG nova.compute.manager [req-0b28bae8-107a-4541-a9e6-1a73d0d63f29 req-a404d142-ee30-4160-adf3-4595a05dba1e service nova] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Refreshing instance network info cache due to event network-changed-1080e2f3-c9a3-48bc-bd65-b5f231bacdc7. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 831.170256] env[61970]: DEBUG oslo_concurrency.lockutils [req-0b28bae8-107a-4541-a9e6-1a73d0d63f29 req-a404d142-ee30-4160-adf3-4595a05dba1e service nova] Acquiring lock "refresh_cache-5d4d97e1-5812-4884-a1ea-6eb0cee591c0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.170489] env[61970]: DEBUG oslo_concurrency.lockutils [req-0b28bae8-107a-4541-a9e6-1a73d0d63f29 req-a404d142-ee30-4160-adf3-4595a05dba1e service nova] Acquired lock "refresh_cache-5d4d97e1-5812-4884-a1ea-6eb0cee591c0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.170670] env[61970]: DEBUG nova.network.neutron [req-0b28bae8-107a-4541-a9e6-1a73d0d63f29 req-a404d142-ee30-4160-adf3-4595a05dba1e service nova] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Refreshing network info cache for port 1080e2f3-c9a3-48bc-bd65-b5f231bacdc7 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 831.208539] env[61970]: DEBUG nova.compute.manager [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 831.250092] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "a78b63a5-6bb8-4271-90d8-1e86fb29db4f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.250380] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "a78b63a5-6bb8-4271-90d8-1e86fb29db4f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.251028] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "a78b63a5-6bb8-4271-90d8-1e86fb29db4f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.251028] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "a78b63a5-6bb8-4271-90d8-1e86fb29db4f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.251028] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "a78b63a5-6bb8-4271-90d8-1e86fb29db4f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.254672] env[61970]: INFO nova.compute.manager [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Terminating instance [ 831.255780] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "refresh_cache-a78b63a5-6bb8-4271-90d8-1e86fb29db4f" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.255780] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquired lock "refresh_cache-a78b63a5-6bb8-4271-90d8-1e86fb29db4f" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.255780] env[61970]: DEBUG nova.network.neutron [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 831.264594] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "refresh_cache-5d4d97e1-5812-4884-a1ea-6eb0cee591c0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.489526] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f66f6f59-bde3-451c-99e6-433cb922d696 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.498868] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32f704cd-2d32-4968-98a7-ddce432a4866 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.529626] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e16d5a2-7863-418c-bf64-47d6f5a9abf0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.536589] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a58a604f-ef3f-4c75-be87-503fd70d4283 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.549836] env[61970]: DEBUG nova.compute.provider_tree [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.702522] env[61970]: DEBUG nova.network.neutron [req-0b28bae8-107a-4541-a9e6-1a73d0d63f29 req-a404d142-ee30-4160-adf3-4595a05dba1e service nova] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 831.769156] env[61970]: DEBUG oslo_concurrency.lockutils [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "0304640a-6c46-4f57-a8b8-5058da2e3de4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.769831] env[61970]: DEBUG oslo_concurrency.lockutils [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "0304640a-6c46-4f57-a8b8-5058da2e3de4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.773946] env[61970]: DEBUG nova.network.neutron [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 831.789769] env[61970]: DEBUG nova.network.neutron [req-0b28bae8-107a-4541-a9e6-1a73d0d63f29 req-a404d142-ee30-4160-adf3-4595a05dba1e service nova] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.820424] env[61970]: DEBUG nova.network.neutron [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.052623] env[61970]: DEBUG nova.scheduler.client.report [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 832.218515] env[61970]: DEBUG nova.compute.manager [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 832.244472] env[61970]: DEBUG nova.virt.hardware [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 832.244706] env[61970]: DEBUG nova.virt.hardware [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 832.244864] env[61970]: DEBUG nova.virt.hardware [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 832.245057] env[61970]: DEBUG nova.virt.hardware [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 832.245208] env[61970]: DEBUG nova.virt.hardware [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 832.245355] env[61970]: DEBUG nova.virt.hardware [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 832.245565] env[61970]: DEBUG nova.virt.hardware [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 832.245719] env[61970]: DEBUG nova.virt.hardware [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 832.245879] env[61970]: DEBUG nova.virt.hardware [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 832.246055] env[61970]: DEBUG nova.virt.hardware [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 832.246233] env[61970]: DEBUG nova.virt.hardware [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.247099] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9996d6b7-b551-43c9-8e07-3d940aa13466 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.255038] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f0305af-f636-42d6-b6a3-9a36948686b7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.292109] env[61970]: DEBUG oslo_concurrency.lockutils [req-0b28bae8-107a-4541-a9e6-1a73d0d63f29 req-a404d142-ee30-4160-adf3-4595a05dba1e service nova] Releasing lock "refresh_cache-5d4d97e1-5812-4884-a1ea-6eb0cee591c0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.292497] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquired lock "refresh_cache-5d4d97e1-5812-4884-a1ea-6eb0cee591c0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.292653] env[61970]: DEBUG nova.network.neutron [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 832.323355] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Releasing lock "refresh_cache-a78b63a5-6bb8-4271-90d8-1e86fb29db4f" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.323754] env[61970]: DEBUG nova.compute.manager [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 832.323939] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 832.324790] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4b44ab7-454c-4173-958b-3e1c73361f7a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.332400] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 832.332643] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e1a67d37-b585-4d1d-a352-e3ca617a5833 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.338365] env[61970]: DEBUG oslo_vmware.api [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 832.338365] env[61970]: value = "task-1355585" [ 832.338365] env[61970]: _type = "Task" [ 832.338365] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.345816] env[61970]: DEBUG oslo_vmware.api [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355585, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.557438] env[61970]: DEBUG oslo_concurrency.lockutils [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.364s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.557960] env[61970]: DEBUG nova.compute.manager [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 832.560504] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.187s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.739285] env[61970]: DEBUG nova.network.neutron [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Successfully updated port: c2860647-20d9-411e-9d79-e26336519408 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 832.821565] env[61970]: DEBUG nova.network.neutron [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.848037] env[61970]: DEBUG oslo_vmware.api [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355585, 'name': PowerOffVM_Task, 'duration_secs': 0.109598} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.848307] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 832.848476] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 832.848715] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8425495c-f73a-456d-a7bf-ca23b6e62c52 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.872344] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 832.872573] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 832.872882] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Deleting the datastore file [datastore1] a78b63a5-6bb8-4271-90d8-1e86fb29db4f {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 832.873306] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-099a9d8a-cb9e-4d58-8d85-618547662d0d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.879607] env[61970]: DEBUG oslo_vmware.api [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 832.879607] env[61970]: value = "task-1355587" [ 832.879607] env[61970]: _type = "Task" [ 832.879607] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.887137] env[61970]: DEBUG oslo_vmware.api [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355587, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.002265] env[61970]: DEBUG nova.network.neutron [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Updating instance_info_cache with network_info: [{"id": "1080e2f3-c9a3-48bc-bd65-b5f231bacdc7", "address": "fa:16:3e:55:7f:ba", "network": {"id": "700300d5-dfe3-4ff7-86ed-43a4f2c15818", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1754634816-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e91895bd10c74f8d9a101274d87cdc71", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ba56987-9dc3-4c76-a4e2-942b05355bdb", "external-id": "nsx-vlan-transportzone-698", "segmentation_id": 698, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1080e2f3-c9", "ovs_interfaceid": "1080e2f3-c9a3-48bc-bd65-b5f231bacdc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.065139] env[61970]: DEBUG nova.compute.utils [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 833.069650] env[61970]: DEBUG nova.compute.manager [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 833.069873] env[61970]: DEBUG nova.network.neutron [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 833.109571] env[61970]: DEBUG nova.policy [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98de02a5c4e14e9889be2fa1637b68d7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e91895bd10c74f8d9a101274d87cdc71', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 833.189366] env[61970]: DEBUG nova.compute.manager [req-f38d200d-2832-4f98-b2c8-eba40ff7f4da req-9b39ba70-90e6-400a-882c-ee6372f2b704 service nova] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Received event network-vif-plugged-c2860647-20d9-411e-9d79-e26336519408 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.189575] env[61970]: DEBUG oslo_concurrency.lockutils [req-f38d200d-2832-4f98-b2c8-eba40ff7f4da req-9b39ba70-90e6-400a-882c-ee6372f2b704 service nova] Acquiring lock "6345d239-fbae-4f3a-9f61-79e10539ff0b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.189801] env[61970]: DEBUG oslo_concurrency.lockutils [req-f38d200d-2832-4f98-b2c8-eba40ff7f4da req-9b39ba70-90e6-400a-882c-ee6372f2b704 service nova] Lock "6345d239-fbae-4f3a-9f61-79e10539ff0b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.189932] env[61970]: DEBUG oslo_concurrency.lockutils [req-f38d200d-2832-4f98-b2c8-eba40ff7f4da req-9b39ba70-90e6-400a-882c-ee6372f2b704 service nova] Lock "6345d239-fbae-4f3a-9f61-79e10539ff0b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.190137] env[61970]: DEBUG nova.compute.manager [req-f38d200d-2832-4f98-b2c8-eba40ff7f4da req-9b39ba70-90e6-400a-882c-ee6372f2b704 service nova] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] No waiting events found dispatching network-vif-plugged-c2860647-20d9-411e-9d79-e26336519408 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 833.190271] env[61970]: WARNING nova.compute.manager [req-f38d200d-2832-4f98-b2c8-eba40ff7f4da req-9b39ba70-90e6-400a-882c-ee6372f2b704 service nova] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Received unexpected event network-vif-plugged-c2860647-20d9-411e-9d79-e26336519408 for instance with vm_state building and task_state spawning. [ 833.190425] env[61970]: DEBUG nova.compute.manager [req-f38d200d-2832-4f98-b2c8-eba40ff7f4da req-9b39ba70-90e6-400a-882c-ee6372f2b704 service nova] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Received event network-changed-c2860647-20d9-411e-9d79-e26336519408 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.190638] env[61970]: DEBUG nova.compute.manager [req-f38d200d-2832-4f98-b2c8-eba40ff7f4da req-9b39ba70-90e6-400a-882c-ee6372f2b704 service nova] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Refreshing instance network info cache due to event network-changed-c2860647-20d9-411e-9d79-e26336519408. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 833.191214] env[61970]: DEBUG oslo_concurrency.lockutils [req-f38d200d-2832-4f98-b2c8-eba40ff7f4da req-9b39ba70-90e6-400a-882c-ee6372f2b704 service nova] Acquiring lock "refresh_cache-6345d239-fbae-4f3a-9f61-79e10539ff0b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.191214] env[61970]: DEBUG oslo_concurrency.lockutils [req-f38d200d-2832-4f98-b2c8-eba40ff7f4da req-9b39ba70-90e6-400a-882c-ee6372f2b704 service nova] Acquired lock "refresh_cache-6345d239-fbae-4f3a-9f61-79e10539ff0b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.191214] env[61970]: DEBUG nova.network.neutron [req-f38d200d-2832-4f98-b2c8-eba40ff7f4da req-9b39ba70-90e6-400a-882c-ee6372f2b704 service nova] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Refreshing network info cache for port c2860647-20d9-411e-9d79-e26336519408 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 833.242359] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "refresh_cache-6345d239-fbae-4f3a-9f61-79e10539ff0b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.327106] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d9d38e2-2c18-4c83-b8c9-c953d093a780 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.343942] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-547de7af-d9a9-4040-ae8d-a249f2a24d96 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.379524] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9248b21-bcc2-422c-b6e5-a46fe3eb2aff {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.389899] env[61970]: DEBUG oslo_vmware.api [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355587, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.112419} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.390684] env[61970]: DEBUG nova.network.neutron [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Successfully created port: 817c117f-93e2-414c-ac36-01fbf658b9d1 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 833.394095] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 833.394325] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 833.394524] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 833.394701] env[61970]: INFO nova.compute.manager [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Took 1.07 seconds to destroy the instance on the hypervisor. [ 833.394936] env[61970]: DEBUG oslo.service.loopingcall [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.395211] env[61970]: DEBUG nova.compute.manager [-] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 833.395318] env[61970]: DEBUG nova.network.neutron [-] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 833.397863] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-446405ef-edae-4041-a383-96f0054c91d0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.411663] env[61970]: DEBUG nova.compute.provider_tree [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.414098] env[61970]: DEBUG nova.network.neutron [-] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.504904] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Releasing lock "refresh_cache-5d4d97e1-5812-4884-a1ea-6eb0cee591c0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.505306] env[61970]: DEBUG nova.compute.manager [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Instance network_info: |[{"id": "1080e2f3-c9a3-48bc-bd65-b5f231bacdc7", "address": "fa:16:3e:55:7f:ba", "network": {"id": "700300d5-dfe3-4ff7-86ed-43a4f2c15818", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1754634816-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e91895bd10c74f8d9a101274d87cdc71", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ba56987-9dc3-4c76-a4e2-942b05355bdb", "external-id": "nsx-vlan-transportzone-698", "segmentation_id": 698, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1080e2f3-c9", "ovs_interfaceid": "1080e2f3-c9a3-48bc-bd65-b5f231bacdc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 833.505835] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:7f:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6ba56987-9dc3-4c76-a4e2-942b05355bdb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1080e2f3-c9a3-48bc-bd65-b5f231bacdc7', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 833.514078] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Creating folder: Project (e91895bd10c74f8d9a101274d87cdc71). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.514363] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1a379cad-1de3-4c4b-abe5-469e1ac9b518 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.526009] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Created folder: Project (e91895bd10c74f8d9a101274d87cdc71) in parent group-v288740. [ 833.526207] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Creating folder: Instances. Parent ref: group-v288770. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.526449] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1004be16-6842-49b6-b041-0e6a0b5bb7aa {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.535700] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Created folder: Instances in parent group-v288770. [ 833.535930] env[61970]: DEBUG oslo.service.loopingcall [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.536134] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 833.536330] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b633a567-ad69-4a3b-afcd-2dd87867e064 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.554961] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 833.554961] env[61970]: value = "task-1355590" [ 833.554961] env[61970]: _type = "Task" [ 833.554961] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.562105] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355590, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.572754] env[61970]: DEBUG nova.compute.manager [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 833.776163] env[61970]: DEBUG nova.network.neutron [req-f38d200d-2832-4f98-b2c8-eba40ff7f4da req-9b39ba70-90e6-400a-882c-ee6372f2b704 service nova] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.879468] env[61970]: DEBUG nova.network.neutron [req-f38d200d-2832-4f98-b2c8-eba40ff7f4da req-9b39ba70-90e6-400a-882c-ee6372f2b704 service nova] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.915071] env[61970]: DEBUG nova.scheduler.client.report [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 833.918598] env[61970]: DEBUG nova.network.neutron [-] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.065852] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355590, 'name': CreateVM_Task, 'duration_secs': 0.34643} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.066299] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 834.072652] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.072882] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.073250] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 834.073569] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed1b901b-e2af-40ab-8d86-2432cf37388f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.081067] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 834.081067] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52686755-5f1b-2c77-d48b-a1d3eb831a6e" [ 834.081067] env[61970]: _type = "Task" [ 834.081067] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.088653] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52686755-5f1b-2c77-d48b-a1d3eb831a6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.382829] env[61970]: DEBUG oslo_concurrency.lockutils [req-f38d200d-2832-4f98-b2c8-eba40ff7f4da req-9b39ba70-90e6-400a-882c-ee6372f2b704 service nova] Releasing lock "refresh_cache-6345d239-fbae-4f3a-9f61-79e10539ff0b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.383112] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquired lock "refresh_cache-6345d239-fbae-4f3a-9f61-79e10539ff0b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.383233] env[61970]: DEBUG nova.network.neutron [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 834.419994] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.859s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.420817] env[61970]: ERROR nova.compute.manager [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9248c5b1-6dfe-4447-a7e4-805e208e2ce3, please check neutron logs for more information. [ 834.420817] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Traceback (most recent call last): [ 834.420817] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 834.420817] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] self.driver.spawn(context, instance, image_meta, [ 834.420817] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 834.420817] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] self._vmops.spawn(context, instance, image_meta, injected_files, [ 834.420817] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 834.420817] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] vm_ref = self.build_virtual_machine(instance, [ 834.420817] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 834.420817] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] vif_infos = vmwarevif.get_vif_info(self._session, [ 834.420817] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 834.421144] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] for vif in network_info: [ 834.421144] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 834.421144] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] return self._sync_wrapper(fn, *args, **kwargs) [ 834.421144] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 834.421144] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] self.wait() [ 834.421144] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 834.421144] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] self[:] = self._gt.wait() [ 834.421144] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 834.421144] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] return self._exit_event.wait() [ 834.421144] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 834.421144] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] current.throw(*self._exc) [ 834.421144] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 834.421144] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] result = function(*args, **kwargs) [ 834.421521] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 834.421521] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] return func(*args, **kwargs) [ 834.421521] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 834.421521] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] raise e [ 834.421521] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 834.421521] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] nwinfo = self.network_api.allocate_for_instance( [ 834.421521] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 834.421521] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] created_port_ids = self._update_ports_for_instance( [ 834.421521] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 834.421521] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] with excutils.save_and_reraise_exception(): [ 834.421521] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 834.421521] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] self.force_reraise() [ 834.421521] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 834.421866] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] raise self.value [ 834.421866] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 834.421866] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] updated_port = self._update_port( [ 834.421866] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 834.421866] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] _ensure_no_port_binding_failure(port) [ 834.421866] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 834.421866] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] raise exception.PortBindingFailed(port_id=port['id']) [ 834.421866] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] nova.exception.PortBindingFailed: Binding failed for port 9248c5b1-6dfe-4447-a7e4-805e208e2ce3, please check neutron logs for more information. [ 834.421866] env[61970]: ERROR nova.compute.manager [instance: 27996960-d2d4-496c-b52d-5d7312aed040] [ 834.422964] env[61970]: DEBUG nova.compute.utils [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Binding failed for port 9248c5b1-6dfe-4447-a7e4-805e208e2ce3, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 834.424126] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.974s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.425693] env[61970]: INFO nova.compute.claims [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 834.428528] env[61970]: INFO nova.compute.manager [-] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Took 1.03 seconds to deallocate network for instance. [ 834.428698] env[61970]: DEBUG nova.compute.manager [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Build of instance 27996960-d2d4-496c-b52d-5d7312aed040 was re-scheduled: Binding failed for port 9248c5b1-6dfe-4447-a7e4-805e208e2ce3, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 834.429144] env[61970]: DEBUG nova.compute.manager [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 834.429368] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Acquiring lock "refresh_cache-27996960-d2d4-496c-b52d-5d7312aed040" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.429515] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Acquired lock "refresh_cache-27996960-d2d4-496c-b52d-5d7312aed040" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.429670] env[61970]: DEBUG nova.network.neutron [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 834.582732] env[61970]: DEBUG nova.compute.manager [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 834.594974] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52686755-5f1b-2c77-d48b-a1d3eb831a6e, 'name': SearchDatastore_Task, 'duration_secs': 0.011348} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.595299] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.595538] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.595794] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.595955] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.596150] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 834.596405] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5a0b5661-f9a5-429f-a2b0-9bc8b43c0d66 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.603861] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 834.604048] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 834.604792] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9dd52d89-8739-49ab-8a70-6c08f4b0e89c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.608677] env[61970]: DEBUG nova.virt.hardware [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 834.608894] env[61970]: DEBUG nova.virt.hardware [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 834.609059] env[61970]: DEBUG nova.virt.hardware [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 834.609245] env[61970]: DEBUG nova.virt.hardware [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 834.609389] env[61970]: DEBUG nova.virt.hardware [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 834.609537] env[61970]: DEBUG nova.virt.hardware [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 834.609764] env[61970]: DEBUG nova.virt.hardware [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 834.609944] env[61970]: DEBUG nova.virt.hardware [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 834.610128] env[61970]: DEBUG nova.virt.hardware [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 834.610319] env[61970]: DEBUG nova.virt.hardware [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 834.610505] env[61970]: DEBUG nova.virt.hardware [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 834.611504] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c71458-7e40-468f-b1e3-922a4267b8d2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.616725] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 834.616725] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d17925-464b-7afc-cd40-f1bd5e7b5e3e" [ 834.616725] env[61970]: _type = "Task" [ 834.616725] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.621956] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8e9d9b-5f2f-47e8-97a9-c81fd4fdc580 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.630166] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d17925-464b-7afc-cd40-f1bd5e7b5e3e, 'name': SearchDatastore_Task, 'duration_secs': 0.008394} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.638044] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b08987bd-bb4e-4958-a09f-a4c85b754f17 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.642263] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 834.642263] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5288f0bb-7794-0c40-87ca-15d5c457cdd9" [ 834.642263] env[61970]: _type = "Task" [ 834.642263] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.649182] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5288f0bb-7794-0c40-87ca-15d5c457cdd9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.936245] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.047815] env[61970]: DEBUG nova.network.neutron [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.061077] env[61970]: DEBUG nova.network.neutron [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Successfully updated port: 817c117f-93e2-414c-ac36-01fbf658b9d1 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 835.124024] env[61970]: DEBUG nova.network.neutron [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.154072] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5288f0bb-7794-0c40-87ca-15d5c457cdd9, 'name': SearchDatastore_Task, 'duration_secs': 0.007898} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.154923] env[61970]: DEBUG nova.network.neutron [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.156657] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.156964] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 5d4d97e1-5812-4884-a1ea-6eb0cee591c0/5d4d97e1-5812-4884-a1ea-6eb0cee591c0.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 835.159771] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8dd4171d-e8fe-40fa-93a8-8a65eed22438 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.166626] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 835.166626] env[61970]: value = "task-1355591" [ 835.166626] env[61970]: _type = "Task" [ 835.166626] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.178728] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355591, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.238133] env[61970]: DEBUG nova.compute.manager [req-675a1b96-7408-4673-af09-45b442702bb1 req-7efcdf39-0c13-4379-97c8-4f1ba8ac4863 service nova] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Received event network-vif-plugged-817c117f-93e2-414c-ac36-01fbf658b9d1 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.238358] env[61970]: DEBUG oslo_concurrency.lockutils [req-675a1b96-7408-4673-af09-45b442702bb1 req-7efcdf39-0c13-4379-97c8-4f1ba8ac4863 service nova] Acquiring lock "e2f51299-56fe-46cc-9e08-2246178b9db7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.238533] env[61970]: DEBUG oslo_concurrency.lockutils [req-675a1b96-7408-4673-af09-45b442702bb1 req-7efcdf39-0c13-4379-97c8-4f1ba8ac4863 service nova] Lock "e2f51299-56fe-46cc-9e08-2246178b9db7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.238700] env[61970]: DEBUG oslo_concurrency.lockutils [req-675a1b96-7408-4673-af09-45b442702bb1 req-7efcdf39-0c13-4379-97c8-4f1ba8ac4863 service nova] Lock "e2f51299-56fe-46cc-9e08-2246178b9db7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.238879] env[61970]: DEBUG nova.compute.manager [req-675a1b96-7408-4673-af09-45b442702bb1 req-7efcdf39-0c13-4379-97c8-4f1ba8ac4863 service nova] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] No waiting events found dispatching network-vif-plugged-817c117f-93e2-414c-ac36-01fbf658b9d1 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 835.239056] env[61970]: WARNING nova.compute.manager [req-675a1b96-7408-4673-af09-45b442702bb1 req-7efcdf39-0c13-4379-97c8-4f1ba8ac4863 service nova] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Received unexpected event network-vif-plugged-817c117f-93e2-414c-ac36-01fbf658b9d1 for instance with vm_state building and task_state spawning. [ 835.239219] env[61970]: DEBUG nova.compute.manager [req-675a1b96-7408-4673-af09-45b442702bb1 req-7efcdf39-0c13-4379-97c8-4f1ba8ac4863 service nova] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Received event network-changed-817c117f-93e2-414c-ac36-01fbf658b9d1 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.239371] env[61970]: DEBUG nova.compute.manager [req-675a1b96-7408-4673-af09-45b442702bb1 req-7efcdf39-0c13-4379-97c8-4f1ba8ac4863 service nova] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Refreshing instance network info cache due to event network-changed-817c117f-93e2-414c-ac36-01fbf658b9d1. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 835.239549] env[61970]: DEBUG oslo_concurrency.lockutils [req-675a1b96-7408-4673-af09-45b442702bb1 req-7efcdf39-0c13-4379-97c8-4f1ba8ac4863 service nova] Acquiring lock "refresh_cache-e2f51299-56fe-46cc-9e08-2246178b9db7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.239678] env[61970]: DEBUG oslo_concurrency.lockutils [req-675a1b96-7408-4673-af09-45b442702bb1 req-7efcdf39-0c13-4379-97c8-4f1ba8ac4863 service nova] Acquired lock "refresh_cache-e2f51299-56fe-46cc-9e08-2246178b9db7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.239831] env[61970]: DEBUG nova.network.neutron [req-675a1b96-7408-4673-af09-45b442702bb1 req-7efcdf39-0c13-4379-97c8-4f1ba8ac4863 service nova] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Refreshing network info cache for port 817c117f-93e2-414c-ac36-01fbf658b9d1 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 835.322184] env[61970]: DEBUG nova.network.neutron [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Updating instance_info_cache with network_info: [{"id": "c2860647-20d9-411e-9d79-e26336519408", "address": "fa:16:3e:7a:23:6d", "network": {"id": "700300d5-dfe3-4ff7-86ed-43a4f2c15818", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1754634816-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e91895bd10c74f8d9a101274d87cdc71", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ba56987-9dc3-4c76-a4e2-942b05355bdb", "external-id": "nsx-vlan-transportzone-698", "segmentation_id": 698, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2860647-20", "ovs_interfaceid": "c2860647-20d9-411e-9d79-e26336519408", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.562013] env[61970]: DEBUG oslo_concurrency.lockutils [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "refresh_cache-e2f51299-56fe-46cc-9e08-2246178b9db7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.658024] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Releasing lock "refresh_cache-27996960-d2d4-496c-b52d-5d7312aed040" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.658236] env[61970]: DEBUG nova.compute.manager [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 835.659021] env[61970]: DEBUG nova.compute.manager [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 835.659021] env[61970]: DEBUG nova.network.neutron [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 835.673772] env[61970]: DEBUG nova.network.neutron [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.680050] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355591, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.479384} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.680289] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 5d4d97e1-5812-4884-a1ea-6eb0cee591c0/5d4d97e1-5812-4884-a1ea-6eb0cee591c0.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 835.680560] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 835.681327] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4b3c83a1-38b0-46ef-a8da-7d41607d852f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.690278] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 835.690278] env[61970]: value = "task-1355592" [ 835.690278] env[61970]: _type = "Task" [ 835.690278] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.701062] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355592, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.707108] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45856ba5-39ed-433f-8d07-336f58ede4f0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.714054] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ea9158f-c8ab-4398-8b18-51e727e9bcb8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.745840] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33612358-1350-452f-b378-405ad668b81c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.753175] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5147408f-d65b-4ba0-9b25-32d166e3ca11 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.766370] env[61970]: DEBUG nova.compute.provider_tree [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.774942] env[61970]: DEBUG nova.network.neutron [req-675a1b96-7408-4673-af09-45b442702bb1 req-7efcdf39-0c13-4379-97c8-4f1ba8ac4863 service nova] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.823730] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Releasing lock "refresh_cache-6345d239-fbae-4f3a-9f61-79e10539ff0b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.824019] env[61970]: DEBUG nova.compute.manager [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Instance network_info: |[{"id": "c2860647-20d9-411e-9d79-e26336519408", "address": "fa:16:3e:7a:23:6d", "network": {"id": "700300d5-dfe3-4ff7-86ed-43a4f2c15818", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1754634816-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e91895bd10c74f8d9a101274d87cdc71", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ba56987-9dc3-4c76-a4e2-942b05355bdb", "external-id": "nsx-vlan-transportzone-698", "segmentation_id": 698, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2860647-20", "ovs_interfaceid": "c2860647-20d9-411e-9d79-e26336519408", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 835.824399] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7a:23:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6ba56987-9dc3-4c76-a4e2-942b05355bdb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2860647-20d9-411e-9d79-e26336519408', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 835.831731] env[61970]: DEBUG oslo.service.loopingcall [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.831935] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 835.832169] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8941c21c-f47d-430e-a6d4-1f44a1b9b1fc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.847243] env[61970]: DEBUG nova.network.neutron [req-675a1b96-7408-4673-af09-45b442702bb1 req-7efcdf39-0c13-4379-97c8-4f1ba8ac4863 service nova] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.854741] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 835.854741] env[61970]: value = "task-1355593" [ 835.854741] env[61970]: _type = "Task" [ 835.854741] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.863798] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355593, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.176342] env[61970]: DEBUG nova.network.neutron [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.200110] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355592, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060621} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.200387] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 836.201170] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c2b0fc-f5ac-4af0-a48d-ae283c6ddf19 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.222814] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] 5d4d97e1-5812-4884-a1ea-6eb0cee591c0/5d4d97e1-5812-4884-a1ea-6eb0cee591c0.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 836.223382] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b52036a4-2350-4d89-bc3b-26e70b212d4f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.241652] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 836.241652] env[61970]: value = "task-1355594" [ 836.241652] env[61970]: _type = "Task" [ 836.241652] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.249398] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355594, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.269224] env[61970]: DEBUG nova.scheduler.client.report [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 836.349898] env[61970]: DEBUG oslo_concurrency.lockutils [req-675a1b96-7408-4673-af09-45b442702bb1 req-7efcdf39-0c13-4379-97c8-4f1ba8ac4863 service nova] Releasing lock "refresh_cache-e2f51299-56fe-46cc-9e08-2246178b9db7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.350327] env[61970]: DEBUG oslo_concurrency.lockutils [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquired lock "refresh_cache-e2f51299-56fe-46cc-9e08-2246178b9db7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.350571] env[61970]: DEBUG nova.network.neutron [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 836.366186] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355593, 'name': CreateVM_Task, 'duration_secs': 0.320412} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.366350] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 836.367013] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.367183] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.367499] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 836.367750] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1850e42e-08b3-43e0-8886-11a9b41a1cff {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.372471] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 836.372471] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52781763-850e-d82e-7455-ed4f80434508" [ 836.372471] env[61970]: _type = "Task" [ 836.372471] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.379862] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52781763-850e-d82e-7455-ed4f80434508, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.679019] env[61970]: INFO nova.compute.manager [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] [instance: 27996960-d2d4-496c-b52d-5d7312aed040] Took 1.02 seconds to deallocate network for instance. [ 836.751529] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355594, 'name': ReconfigVM_Task, 'duration_secs': 0.259338} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.751828] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Reconfigured VM instance instance-0000003d to attach disk [datastore2] 5d4d97e1-5812-4884-a1ea-6eb0cee591c0/5d4d97e1-5812-4884-a1ea-6eb0cee591c0.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.752467] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-78fd4a08-7950-419e-a080-6efda284199a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.758607] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 836.758607] env[61970]: value = "task-1355595" [ 836.758607] env[61970]: _type = "Task" [ 836.758607] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.766237] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355595, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.774038] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.774564] env[61970]: DEBUG nova.compute.manager [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 836.777861] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.732s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.779283] env[61970]: INFO nova.compute.claims [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.883109] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52781763-850e-d82e-7455-ed4f80434508, 'name': SearchDatastore_Task, 'duration_secs': 0.015338} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.883888] env[61970]: DEBUG nova.network.neutron [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 836.885683] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.886046] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 836.886163] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.886307] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.886483] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.886756] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e85274fc-1aa4-4d38-82e1-9fcec78fe3c5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.894515] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.894690] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 836.895392] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3eb2a8e-7958-44b7-84a9-61570371f741 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.899936] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 836.899936] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]527a2942-07fe-be8b-999d-af602bbd7ded" [ 836.899936] env[61970]: _type = "Task" [ 836.899936] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.908852] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]527a2942-07fe-be8b-999d-af602bbd7ded, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.025758] env[61970]: DEBUG nova.network.neutron [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Updating instance_info_cache with network_info: [{"id": "817c117f-93e2-414c-ac36-01fbf658b9d1", "address": "fa:16:3e:78:5e:7a", "network": {"id": "700300d5-dfe3-4ff7-86ed-43a4f2c15818", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1754634816-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e91895bd10c74f8d9a101274d87cdc71", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ba56987-9dc3-4c76-a4e2-942b05355bdb", "external-id": "nsx-vlan-transportzone-698", "segmentation_id": 698, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap817c117f-93", "ovs_interfaceid": "817c117f-93e2-414c-ac36-01fbf658b9d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.268452] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355595, 'name': Rename_Task, 'duration_secs': 0.12189} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.268761] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 837.269013] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8bf80f9d-7d34-4bab-8160-bcd0b6e30812 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.274877] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 837.274877] env[61970]: value = "task-1355596" [ 837.274877] env[61970]: _type = "Task" [ 837.274877] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.278587] env[61970]: DEBUG nova.compute.utils [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 837.280174] env[61970]: DEBUG nova.compute.manager [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 837.280784] env[61970]: DEBUG nova.network.neutron [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 837.288387] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355596, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.350115] env[61970]: DEBUG nova.policy [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd04344afddb14418b5018b69fbd7694e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c6a4997482a4834a30c79c4cbab2cc1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 837.412686] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]527a2942-07fe-be8b-999d-af602bbd7ded, 'name': SearchDatastore_Task, 'duration_secs': 0.012494} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.413793] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f3eeaeb-3c6a-430f-8069-2ee855cf780c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.418980] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 837.418980] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f47a31-fc5e-0bde-2e16-a158039dcf86" [ 837.418980] env[61970]: _type = "Task" [ 837.418980] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.427184] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f47a31-fc5e-0bde-2e16-a158039dcf86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.529029] env[61970]: DEBUG oslo_concurrency.lockutils [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Releasing lock "refresh_cache-e2f51299-56fe-46cc-9e08-2246178b9db7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.530195] env[61970]: DEBUG nova.compute.manager [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Instance network_info: |[{"id": "817c117f-93e2-414c-ac36-01fbf658b9d1", "address": "fa:16:3e:78:5e:7a", "network": {"id": "700300d5-dfe3-4ff7-86ed-43a4f2c15818", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1754634816-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e91895bd10c74f8d9a101274d87cdc71", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ba56987-9dc3-4c76-a4e2-942b05355bdb", "external-id": "nsx-vlan-transportzone-698", "segmentation_id": 698, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap817c117f-93", "ovs_interfaceid": "817c117f-93e2-414c-ac36-01fbf658b9d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 837.530332] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:5e:7a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6ba56987-9dc3-4c76-a4e2-942b05355bdb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '817c117f-93e2-414c-ac36-01fbf658b9d1', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.537700] env[61970]: DEBUG oslo.service.loopingcall [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.539363] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 837.539363] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-67267558-a780-452d-b89c-45f04340f9d8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.557625] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.557625] env[61970]: value = "task-1355597" [ 837.557625] env[61970]: _type = "Task" [ 837.557625] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.567259] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355597, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.677782] env[61970]: DEBUG nova.network.neutron [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Successfully created port: 90ade7be-34cd-42db-b38c-0e64cba25ec4 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 837.709396] env[61970]: INFO nova.scheduler.client.report [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Deleted allocations for instance 27996960-d2d4-496c-b52d-5d7312aed040 [ 837.788553] env[61970]: DEBUG nova.compute.manager [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 837.788553] env[61970]: DEBUG oslo_vmware.api [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355596, 'name': PowerOnVM_Task, 'duration_secs': 0.448886} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.791506] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 837.791769] env[61970]: INFO nova.compute.manager [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Took 7.99 seconds to spawn the instance on the hypervisor. [ 837.791953] env[61970]: DEBUG nova.compute.manager [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.793180] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79820f57-dec6-4e7f-ae2a-330fc4139110 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.929752] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f47a31-fc5e-0bde-2e16-a158039dcf86, 'name': SearchDatastore_Task, 'duration_secs': 0.009733} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.930171] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.930171] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 6345d239-fbae-4f3a-9f61-79e10539ff0b/6345d239-fbae-4f3a-9f61-79e10539ff0b.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 837.930545] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ebbc7ed0-79a3-404e-b46a-6de8b26e26cc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.936893] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 837.936893] env[61970]: value = "task-1355598" [ 837.936893] env[61970]: _type = "Task" [ 837.936893] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.946819] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355598, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.068012] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355597, 'name': CreateVM_Task, 'duration_secs': 0.314487} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.068265] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 838.069041] env[61970]: DEBUG oslo_concurrency.lockutils [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.069282] env[61970]: DEBUG oslo_concurrency.lockutils [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.069630] env[61970]: DEBUG oslo_concurrency.lockutils [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 838.070668] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52406ae6-f8b1-4f55-9487-decc529b8ac4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.072983] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9258d6e6-f2cd-4088-93ed-a72f5ade2480 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.079611] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 838.079611] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529e9257-90f8-3b92-b298-751a0e78a3f3" [ 838.079611] env[61970]: _type = "Task" [ 838.079611] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.084424] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-345e0928-67bc-4123-8d87-b278cf7e24e4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.092388] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529e9257-90f8-3b92-b298-751a0e78a3f3, 'name': SearchDatastore_Task, 'duration_secs': 0.008225} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.123126] env[61970]: DEBUG oslo_concurrency.lockutils [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.123492] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.123788] env[61970]: DEBUG oslo_concurrency.lockutils [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.123981] env[61970]: DEBUG oslo_concurrency.lockutils [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.124225] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.125153] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e8053d51-4bc3-44e9-861b-9449e153553c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.127694] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed7559c-db00-4bf5-9c61-60400c7e17f9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.137610] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a719d14-8afe-494c-a362-d3e71206c235 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.143776] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.144139] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 838.145521] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ae11ca1-dd4f-4282-8a88-20086250a777 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.157403] env[61970]: DEBUG nova.compute.provider_tree [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.161143] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 838.161143] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5287551a-0c55-3010-46af-58497f3747db" [ 838.161143] env[61970]: _type = "Task" [ 838.161143] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.169816] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5287551a-0c55-3010-46af-58497f3747db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.218539] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ba419a2f-ad27-4278-906b-2e9ae2d98174 tempest-ServerPasswordTestJSON-1050340498 tempest-ServerPasswordTestJSON-1050340498-project-member] Lock "27996960-d2d4-496c-b52d-5d7312aed040" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.211s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.313069] env[61970]: INFO nova.compute.manager [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Took 24.86 seconds to build instance. [ 838.447426] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355598, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473061} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.447644] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 6345d239-fbae-4f3a-9f61-79e10539ff0b/6345d239-fbae-4f3a-9f61-79e10539ff0b.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 838.447857] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 838.448118] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a948d00-3947-455c-ac25-bb7ff5f4dde5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.455057] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 838.455057] env[61970]: value = "task-1355599" [ 838.455057] env[61970]: _type = "Task" [ 838.455057] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.462784] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355599, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.660742] env[61970]: DEBUG nova.scheduler.client.report [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.678030] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5287551a-0c55-3010-46af-58497f3747db, 'name': SearchDatastore_Task, 'duration_secs': 0.053992} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.679051] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d7b99ad-8801-42c0-8369-c100d2855af3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.686118] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 838.686118] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5218fec3-9bbb-0389-e7ed-9fa270e114b6" [ 838.686118] env[61970]: _type = "Task" [ 838.686118] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.698123] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5218fec3-9bbb-0389-e7ed-9fa270e114b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.722082] env[61970]: DEBUG nova.compute.manager [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 838.800045] env[61970]: DEBUG nova.compute.manager [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 838.815672] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7cb48e47-790c-4dc5-a667-665ea0625738 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "5d4d97e1-5812-4884-a1ea-6eb0cee591c0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.323s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.829460] env[61970]: DEBUG nova.virt.hardware [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 838.829460] env[61970]: DEBUG nova.virt.hardware [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 838.829460] env[61970]: DEBUG nova.virt.hardware [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 838.829753] env[61970]: DEBUG nova.virt.hardware [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 838.829753] env[61970]: DEBUG nova.virt.hardware [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 838.829753] env[61970]: DEBUG nova.virt.hardware [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 838.829753] env[61970]: DEBUG nova.virt.hardware [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 838.829753] env[61970]: DEBUG nova.virt.hardware [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 838.829908] env[61970]: DEBUG nova.virt.hardware [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 838.829908] env[61970]: DEBUG nova.virt.hardware [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 838.829908] env[61970]: DEBUG nova.virt.hardware [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 838.831071] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b574aff-ae5f-4de9-9938-295d65b6b0b4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.839942] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f3b1c7b-c4bd-4b7b-a220-a258eca81ee1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.965101] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355599, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081315} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.965438] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.969094] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ad59a3-9134-471e-b55b-b53314bc6b1d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.990020] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 6345d239-fbae-4f3a-9f61-79e10539ff0b/6345d239-fbae-4f3a-9f61-79e10539ff0b.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.990020] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-afbf1fed-5736-4638-a7e8-fed7e955a8f1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.007806] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 839.007806] env[61970]: value = "task-1355600" [ 839.007806] env[61970]: _type = "Task" [ 839.007806] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.015449] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355600, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.172471] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.172986] env[61970]: DEBUG nova.compute.manager [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 839.176255] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.171s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.178384] env[61970]: INFO nova.compute.claims [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 839.200433] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5218fec3-9bbb-0389-e7ed-9fa270e114b6, 'name': SearchDatastore_Task, 'duration_secs': 0.032589} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.200433] env[61970]: DEBUG oslo_concurrency.lockutils [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.200433] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] e2f51299-56fe-46cc-9e08-2246178b9db7/e2f51299-56fe-46cc-9e08-2246178b9db7.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 839.200756] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-de659acf-fa08-4c83-8647-11912310ce59 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.210029] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 839.210029] env[61970]: value = "task-1355601" [ 839.210029] env[61970]: _type = "Task" [ 839.210029] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.218269] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355601, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.246456] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.257964] env[61970]: DEBUG nova.compute.manager [req-7db6d0c1-fba2-4625-a81b-6bb3a9057af2 req-6f712ab4-28a6-4833-985c-d605fffcc46c service nova] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Received event network-vif-plugged-90ade7be-34cd-42db-b38c-0e64cba25ec4 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 839.258257] env[61970]: DEBUG oslo_concurrency.lockutils [req-7db6d0c1-fba2-4625-a81b-6bb3a9057af2 req-6f712ab4-28a6-4833-985c-d605fffcc46c service nova] Acquiring lock "b19cb516-b163-4bed-ba5b-139a0a18fc05-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.258466] env[61970]: DEBUG oslo_concurrency.lockutils [req-7db6d0c1-fba2-4625-a81b-6bb3a9057af2 req-6f712ab4-28a6-4833-985c-d605fffcc46c service nova] Lock "b19cb516-b163-4bed-ba5b-139a0a18fc05-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.258636] env[61970]: DEBUG oslo_concurrency.lockutils [req-7db6d0c1-fba2-4625-a81b-6bb3a9057af2 req-6f712ab4-28a6-4833-985c-d605fffcc46c service nova] Lock "b19cb516-b163-4bed-ba5b-139a0a18fc05-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.258801] env[61970]: DEBUG nova.compute.manager [req-7db6d0c1-fba2-4625-a81b-6bb3a9057af2 req-6f712ab4-28a6-4833-985c-d605fffcc46c service nova] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] No waiting events found dispatching network-vif-plugged-90ade7be-34cd-42db-b38c-0e64cba25ec4 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 839.258956] env[61970]: WARNING nova.compute.manager [req-7db6d0c1-fba2-4625-a81b-6bb3a9057af2 req-6f712ab4-28a6-4833-985c-d605fffcc46c service nova] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Received unexpected event network-vif-plugged-90ade7be-34cd-42db-b38c-0e64cba25ec4 for instance with vm_state building and task_state spawning. [ 839.319536] env[61970]: DEBUG nova.compute.manager [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 839.520028] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355600, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.685546] env[61970]: DEBUG nova.compute.utils [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 839.688526] env[61970]: DEBUG nova.compute.manager [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 839.688932] env[61970]: DEBUG nova.network.neutron [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 839.722750] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355601, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.501406} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.722750] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] e2f51299-56fe-46cc-9e08-2246178b9db7/e2f51299-56fe-46cc-9e08-2246178b9db7.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 839.722750] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 839.725565] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ab5237c-a2ed-48c4-957d-b6297c116bff {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.733019] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 839.733019] env[61970]: value = "task-1355602" [ 839.733019] env[61970]: _type = "Task" [ 839.733019] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.743031] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355602, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.786307] env[61970]: DEBUG nova.policy [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2af915dc43c84940a937dfb8a04ea0b6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17ca05a1e6664430a02de563d98c1148', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 839.787696] env[61970]: DEBUG nova.network.neutron [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Successfully updated port: 90ade7be-34cd-42db-b38c-0e64cba25ec4 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 839.843275] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.023190] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355600, 'name': ReconfigVM_Task, 'duration_secs': 0.645841} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.025702] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 6345d239-fbae-4f3a-9f61-79e10539ff0b/6345d239-fbae-4f3a-9f61-79e10539ff0b.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 840.025702] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3d95f519-67fb-44a8-9cbe-a5daeb6d9fff {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.031571] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 840.031571] env[61970]: value = "task-1355603" [ 840.031571] env[61970]: _type = "Task" [ 840.031571] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.040568] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355603, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.105012] env[61970]: DEBUG nova.compute.manager [req-d8894b50-785e-4b41-8d84-73799bba3e17 req-8a4389d4-5a37-45d9-9260-5978c171c0bd service nova] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Received event network-changed-90ade7be-34cd-42db-b38c-0e64cba25ec4 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.105225] env[61970]: DEBUG nova.compute.manager [req-d8894b50-785e-4b41-8d84-73799bba3e17 req-8a4389d4-5a37-45d9-9260-5978c171c0bd service nova] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Refreshing instance network info cache due to event network-changed-90ade7be-34cd-42db-b38c-0e64cba25ec4. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 840.105440] env[61970]: DEBUG oslo_concurrency.lockutils [req-d8894b50-785e-4b41-8d84-73799bba3e17 req-8a4389d4-5a37-45d9-9260-5978c171c0bd service nova] Acquiring lock "refresh_cache-b19cb516-b163-4bed-ba5b-139a0a18fc05" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.105627] env[61970]: DEBUG oslo_concurrency.lockutils [req-d8894b50-785e-4b41-8d84-73799bba3e17 req-8a4389d4-5a37-45d9-9260-5978c171c0bd service nova] Acquired lock "refresh_cache-b19cb516-b163-4bed-ba5b-139a0a18fc05" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.105811] env[61970]: DEBUG nova.network.neutron [req-d8894b50-785e-4b41-8d84-73799bba3e17 req-8a4389d4-5a37-45d9-9260-5978c171c0bd service nova] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Refreshing network info cache for port 90ade7be-34cd-42db-b38c-0e64cba25ec4 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 840.193461] env[61970]: DEBUG nova.compute.manager [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 840.198840] env[61970]: DEBUG nova.network.neutron [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Successfully created port: 64a3e684-f5d3-4dd2-b8c9-dc46565f71bd {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 840.243260] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355602, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063378} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.243633] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 840.244812] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d496dc-42a8-42b4-95db-7eea7f8f72b7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.269503] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] e2f51299-56fe-46cc-9e08-2246178b9db7/e2f51299-56fe-46cc-9e08-2246178b9db7.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.273479] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6b196db-93c6-499b-a842-89cb21a12237 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.291346] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "refresh_cache-b19cb516-b163-4bed-ba5b-139a0a18fc05" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.296204] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 840.296204] env[61970]: value = "task-1355604" [ 840.296204] env[61970]: _type = "Task" [ 840.296204] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.302349] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355604, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.507197] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9f7cb8-ad41-497e-9900-d5f50d35bda1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.514930] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-986e3a05-82f2-49a8-8098-4ba2512c5c0e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.549513] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d8b6dd-2aaf-4abc-9d08-a32d25149199 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.558137] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355603, 'name': Rename_Task, 'duration_secs': 0.276829} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.560396] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 840.560741] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c834571d-babe-4ddd-af77-b27fc4409841 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.564245] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa1e7cc-b5e5-4d59-aff9-de3762de011b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.577785] env[61970]: DEBUG nova.compute.provider_tree [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.580299] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 840.580299] env[61970]: value = "task-1355605" [ 840.580299] env[61970]: _type = "Task" [ 840.580299] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.588796] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355605, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.637538] env[61970]: DEBUG nova.network.neutron [req-d8894b50-785e-4b41-8d84-73799bba3e17 req-8a4389d4-5a37-45d9-9260-5978c171c0bd service nova] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 840.713031] env[61970]: DEBUG nova.network.neutron [req-d8894b50-785e-4b41-8d84-73799bba3e17 req-8a4389d4-5a37-45d9-9260-5978c171c0bd service nova] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.803739] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355604, 'name': ReconfigVM_Task, 'duration_secs': 0.287451} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.804074] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Reconfigured VM instance instance-0000003f to attach disk [datastore2] e2f51299-56fe-46cc-9e08-2246178b9db7/e2f51299-56fe-46cc-9e08-2246178b9db7.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 840.804672] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-50be4a9c-099f-484e-a144-cd8a1247ab54 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.810662] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 840.810662] env[61970]: value = "task-1355606" [ 840.810662] env[61970]: _type = "Task" [ 840.810662] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.818556] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355606, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.083128] env[61970]: DEBUG nova.scheduler.client.report [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 841.095820] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355605, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.208482] env[61970]: DEBUG nova.compute.manager [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 841.215135] env[61970]: DEBUG oslo_concurrency.lockutils [req-d8894b50-785e-4b41-8d84-73799bba3e17 req-8a4389d4-5a37-45d9-9260-5978c171c0bd service nova] Releasing lock "refresh_cache-b19cb516-b163-4bed-ba5b-139a0a18fc05" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.215424] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "refresh_cache-b19cb516-b163-4bed-ba5b-139a0a18fc05" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.215598] env[61970]: DEBUG nova.network.neutron [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 841.234600] env[61970]: DEBUG nova.virt.hardware [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 841.234825] env[61970]: DEBUG nova.virt.hardware [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 841.234984] env[61970]: DEBUG nova.virt.hardware [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 841.235181] env[61970]: DEBUG nova.virt.hardware [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 841.235327] env[61970]: DEBUG nova.virt.hardware [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 841.235473] env[61970]: DEBUG nova.virt.hardware [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 841.235679] env[61970]: DEBUG nova.virt.hardware [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 841.235842] env[61970]: DEBUG nova.virt.hardware [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 841.236033] env[61970]: DEBUG nova.virt.hardware [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 841.236199] env[61970]: DEBUG nova.virt.hardware [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 841.236369] env[61970]: DEBUG nova.virt.hardware [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 841.237432] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be35682e-4d8f-449a-b794-74fd9d6eee58 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.244922] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beb24593-0700-4992-8aa7-256120337f00 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.319758] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355606, 'name': Rename_Task, 'duration_secs': 0.136826} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.320019] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 841.320255] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a205d9dc-04f3-499f-b60c-f3222241b336 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.326533] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 841.326533] env[61970]: value = "task-1355607" [ 841.326533] env[61970]: _type = "Task" [ 841.326533] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.335498] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355607, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.591430] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.415s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.591926] env[61970]: DEBUG nova.compute.manager [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 841.595012] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.706s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.596964] env[61970]: INFO nova.compute.claims [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 841.610024] env[61970]: DEBUG oslo_vmware.api [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355605, 'name': PowerOnVM_Task, 'duration_secs': 0.875844} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.610024] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 841.610024] env[61970]: INFO nova.compute.manager [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Took 9.39 seconds to spawn the instance on the hypervisor. [ 841.610024] env[61970]: DEBUG nova.compute.manager [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.610024] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9906203-4df4-488c-a117-af633b9a6d3f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.755519] env[61970]: DEBUG nova.network.neutron [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 841.771838] env[61970]: DEBUG nova.compute.manager [req-26e609a1-1f35-4462-ab21-3b4d96743d51 req-b45544b6-9872-4bf7-a4ca-a91ddfc94890 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Received event network-vif-plugged-64a3e684-f5d3-4dd2-b8c9-dc46565f71bd {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 841.772100] env[61970]: DEBUG oslo_concurrency.lockutils [req-26e609a1-1f35-4462-ab21-3b4d96743d51 req-b45544b6-9872-4bf7-a4ca-a91ddfc94890 service nova] Acquiring lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.772319] env[61970]: DEBUG oslo_concurrency.lockutils [req-26e609a1-1f35-4462-ab21-3b4d96743d51 req-b45544b6-9872-4bf7-a4ca-a91ddfc94890 service nova] Lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.772489] env[61970]: DEBUG oslo_concurrency.lockutils [req-26e609a1-1f35-4462-ab21-3b4d96743d51 req-b45544b6-9872-4bf7-a4ca-a91ddfc94890 service nova] Lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.772653] env[61970]: DEBUG nova.compute.manager [req-26e609a1-1f35-4462-ab21-3b4d96743d51 req-b45544b6-9872-4bf7-a4ca-a91ddfc94890 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] No waiting events found dispatching network-vif-plugged-64a3e684-f5d3-4dd2-b8c9-dc46565f71bd {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 841.772818] env[61970]: WARNING nova.compute.manager [req-26e609a1-1f35-4462-ab21-3b4d96743d51 req-b45544b6-9872-4bf7-a4ca-a91ddfc94890 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Received unexpected event network-vif-plugged-64a3e684-f5d3-4dd2-b8c9-dc46565f71bd for instance with vm_state building and task_state spawning. [ 841.788488] env[61970]: DEBUG nova.network.neutron [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Successfully updated port: 64a3e684-f5d3-4dd2-b8c9-dc46565f71bd {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 841.836245] env[61970]: DEBUG oslo_vmware.api [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355607, 'name': PowerOnVM_Task, 'duration_secs': 0.46068} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.836549] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 841.836587] env[61970]: INFO nova.compute.manager [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Took 7.25 seconds to spawn the instance on the hypervisor. [ 841.836749] env[61970]: DEBUG nova.compute.manager [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.837509] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ce1413-34f4-40d3-b6ce-d4b5e3110f2f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.889806] env[61970]: DEBUG nova.network.neutron [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Updating instance_info_cache with network_info: [{"id": "90ade7be-34cd-42db-b38c-0e64cba25ec4", "address": "fa:16:3e:da:15:38", "network": {"id": "62edd125-b77c-4418-a465-1dcb3a0733c1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-371307366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6a4997482a4834a30c79c4cbab2cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90ade7be-34", "ovs_interfaceid": "90ade7be-34cd-42db-b38c-0e64cba25ec4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.103081] env[61970]: DEBUG nova.compute.utils [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 842.109062] env[61970]: DEBUG nova.compute.manager [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 842.109062] env[61970]: DEBUG nova.network.neutron [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 842.124961] env[61970]: INFO nova.compute.manager [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Took 26.90 seconds to build instance. [ 842.175397] env[61970]: DEBUG nova.policy [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dad53dc4047245d4961960a07e81a5f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2c5df141b77042c28368dc4474f7d917', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 842.291199] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.291360] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.291556] env[61970]: DEBUG nova.network.neutron [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 842.353833] env[61970]: INFO nova.compute.manager [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Took 24.71 seconds to build instance. [ 842.393101] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "refresh_cache-b19cb516-b163-4bed-ba5b-139a0a18fc05" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.393425] env[61970]: DEBUG nova.compute.manager [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Instance network_info: |[{"id": "90ade7be-34cd-42db-b38c-0e64cba25ec4", "address": "fa:16:3e:da:15:38", "network": {"id": "62edd125-b77c-4418-a465-1dcb3a0733c1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-371307366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6a4997482a4834a30c79c4cbab2cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90ade7be-34", "ovs_interfaceid": "90ade7be-34cd-42db-b38c-0e64cba25ec4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 842.393946] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:da:15:38', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '90ade7be-34cd-42db-b38c-0e64cba25ec4', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 842.401528] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Creating folder: Project (9c6a4997482a4834a30c79c4cbab2cc1). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 842.402209] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6bedce56-1202-4018-b41e-32bdd9254ed4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.413307] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Created folder: Project (9c6a4997482a4834a30c79c4cbab2cc1) in parent group-v288740. [ 842.413493] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Creating folder: Instances. Parent ref: group-v288775. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 842.413727] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7ea432f1-1382-4159-bbd6-e338f17170b0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.425366] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Created folder: Instances in parent group-v288775. [ 842.425601] env[61970]: DEBUG oslo.service.loopingcall [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.425782] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 842.425995] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3c12da77-d710-4e67-9889-2b2de6d77c88 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.447384] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 842.447384] env[61970]: value = "task-1355610" [ 842.447384] env[61970]: _type = "Task" [ 842.447384] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.455427] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355610, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.531020] env[61970]: DEBUG nova.network.neutron [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Successfully created port: 6875c46d-b7c5-4912-a1f0-932b3ab900d8 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 842.608597] env[61970]: DEBUG nova.compute.manager [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 842.627532] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de6b6383-2d01-407c-92c4-04969463d7c7 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "6345d239-fbae-4f3a-9f61-79e10539ff0b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.906s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.824042] env[61970]: DEBUG nova.network.neutron [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 842.856517] env[61970]: DEBUG oslo_concurrency.lockutils [None req-92309722-d6d7-4320-acdd-9c0738278c28 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "e2f51299-56fe-46cc-9e08-2246178b9db7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.783s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.906202] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec045d41-3e4d-46fc-b455-63f95ead99c2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.914556] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae72230a-3ab1-459e-9f35-e04bf4ef8b05 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.948391] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51be11c1-b944-4ac4-afb6-6e70af2bafc8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.960938] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8730bb-7206-44a2-9f4a-77c28c6732fc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.964779] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355610, 'name': CreateVM_Task, 'duration_secs': 0.461384} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.964898] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 842.965950] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.966142] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.966454] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 842.966716] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3f3508c-7be9-44e0-8ea7-8252d59029b5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.977784] env[61970]: DEBUG nova.compute.provider_tree [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.981936] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 842.981936] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52476c43-83a1-a93c-70c0-7e05186e6b4b" [ 842.981936] env[61970]: _type = "Task" [ 842.981936] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.991912] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52476c43-83a1-a93c-70c0-7e05186e6b4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.023630] env[61970]: DEBUG nova.network.neutron [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Updating instance_info_cache with network_info: [{"id": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "address": "fa:16:3e:14:8c:ec", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64a3e684-f5", "ovs_interfaceid": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.130196] env[61970]: DEBUG nova.compute.manager [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.362629] env[61970]: DEBUG nova.compute.manager [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.480886] env[61970]: DEBUG nova.scheduler.client.report [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 843.493218] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52476c43-83a1-a93c-70c0-7e05186e6b4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.526810] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.527176] env[61970]: DEBUG nova.compute.manager [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Instance network_info: |[{"id": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "address": "fa:16:3e:14:8c:ec", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64a3e684-f5", "ovs_interfaceid": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 843.527580] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:8c:ec', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '68add7d6-c025-46fa-84d3-9c589adb63e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '64a3e684-f5d3-4dd2-b8c9-dc46565f71bd', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.535433] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Creating folder: Project (17ca05a1e6664430a02de563d98c1148). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 843.535992] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-732a938b-3c68-4138-8e1b-0d75a9a7fed2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.545718] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Created folder: Project (17ca05a1e6664430a02de563d98c1148) in parent group-v288740. [ 843.545894] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Creating folder: Instances. Parent ref: group-v288778. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 843.546151] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ff76e060-7b04-4bdc-8e9c-7a3049d4ee74 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.554672] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Created folder: Instances in parent group-v288778. [ 843.554937] env[61970]: DEBUG oslo.service.loopingcall [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.555140] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 843.555333] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e0afa985-0495-436a-b1c5-6180fe76026e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.574929] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 843.574929] env[61970]: value = "task-1355613" [ 843.574929] env[61970]: _type = "Task" [ 843.574929] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.581952] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355613, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.621614] env[61970]: DEBUG nova.compute.manager [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 843.648252] env[61970]: DEBUG nova.virt.hardware [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 843.648566] env[61970]: DEBUG nova.virt.hardware [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 843.648789] env[61970]: DEBUG nova.virt.hardware [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 843.649454] env[61970]: DEBUG nova.virt.hardware [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 843.649454] env[61970]: DEBUG nova.virt.hardware [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 843.649454] env[61970]: DEBUG nova.virt.hardware [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 843.649607] env[61970]: DEBUG nova.virt.hardware [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 843.649783] env[61970]: DEBUG nova.virt.hardware [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 843.650853] env[61970]: DEBUG nova.virt.hardware [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 843.650853] env[61970]: DEBUG nova.virt.hardware [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 843.650853] env[61970]: DEBUG nova.virt.hardware [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 843.651505] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd27985d-c8d1-4f39-9417-1f17895b1562 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.655491] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.661187] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a01e84-d9e5-47b9-a094-feb267a210c0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.800170] env[61970]: DEBUG nova.compute.manager [req-dcf8d03f-d7be-4e83-a42d-6592e2bd1d28 req-5d3c571e-15cf-4db9-9c60-9be269eb5774 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Received event network-changed-64a3e684-f5d3-4dd2-b8c9-dc46565f71bd {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 843.800170] env[61970]: DEBUG nova.compute.manager [req-dcf8d03f-d7be-4e83-a42d-6592e2bd1d28 req-5d3c571e-15cf-4db9-9c60-9be269eb5774 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Refreshing instance network info cache due to event network-changed-64a3e684-f5d3-4dd2-b8c9-dc46565f71bd. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 843.800170] env[61970]: DEBUG oslo_concurrency.lockutils [req-dcf8d03f-d7be-4e83-a42d-6592e2bd1d28 req-5d3c571e-15cf-4db9-9c60-9be269eb5774 service nova] Acquiring lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.800388] env[61970]: DEBUG oslo_concurrency.lockutils [req-dcf8d03f-d7be-4e83-a42d-6592e2bd1d28 req-5d3c571e-15cf-4db9-9c60-9be269eb5774 service nova] Acquired lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.800681] env[61970]: DEBUG nova.network.neutron [req-dcf8d03f-d7be-4e83-a42d-6592e2bd1d28 req-5d3c571e-15cf-4db9-9c60-9be269eb5774 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Refreshing network info cache for port 64a3e684-f5d3-4dd2-b8c9-dc46565f71bd {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 843.883513] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.988646] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.394s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.989023] env[61970]: DEBUG nova.compute.manager [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 843.992335] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.171s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.993746] env[61970]: INFO nova.compute.claims [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 844.001972] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52476c43-83a1-a93c-70c0-7e05186e6b4b, 'name': SearchDatastore_Task, 'duration_secs': 0.982894} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.002486] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.002733] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.002963] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.003114] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.003287] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.003843] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cbf81bb7-610a-4889-a10a-42aa896ff50a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.083873] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355613, 'name': CreateVM_Task, 'duration_secs': 0.30704} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.087018] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 844.087018] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.087018] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.087018] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 844.087018] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce3534ae-c450-4529-ace3-86e5e32eaba6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.090366] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 844.090366] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]522b7d17-7c4a-5461-155e-7efd10d652b5" [ 844.090366] env[61970]: _type = "Task" [ 844.090366] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.097988] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]522b7d17-7c4a-5461-155e-7efd10d652b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.158171] env[61970]: DEBUG nova.compute.manager [req-c1dee272-be77-4475-b98b-4c6960cdd236 req-5319993b-34c7-447f-9b2c-d6ed75fd05fd service nova] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Received event network-vif-plugged-6875c46d-b7c5-4912-a1f0-932b3ab900d8 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.158171] env[61970]: DEBUG oslo_concurrency.lockutils [req-c1dee272-be77-4475-b98b-4c6960cdd236 req-5319993b-34c7-447f-9b2c-d6ed75fd05fd service nova] Acquiring lock "b385407b-1bdd-4c53-907c-cb4c8ce16cc7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.158171] env[61970]: DEBUG oslo_concurrency.lockutils [req-c1dee272-be77-4475-b98b-4c6960cdd236 req-5319993b-34c7-447f-9b2c-d6ed75fd05fd service nova] Lock "b385407b-1bdd-4c53-907c-cb4c8ce16cc7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.158490] env[61970]: DEBUG oslo_concurrency.lockutils [req-c1dee272-be77-4475-b98b-4c6960cdd236 req-5319993b-34c7-447f-9b2c-d6ed75fd05fd service nova] Lock "b385407b-1bdd-4c53-907c-cb4c8ce16cc7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.158776] env[61970]: DEBUG nova.compute.manager [req-c1dee272-be77-4475-b98b-4c6960cdd236 req-5319993b-34c7-447f-9b2c-d6ed75fd05fd service nova] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] No waiting events found dispatching network-vif-plugged-6875c46d-b7c5-4912-a1f0-932b3ab900d8 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 844.159111] env[61970]: WARNING nova.compute.manager [req-c1dee272-be77-4475-b98b-4c6960cdd236 req-5319993b-34c7-447f-9b2c-d6ed75fd05fd service nova] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Received unexpected event network-vif-plugged-6875c46d-b7c5-4912-a1f0-932b3ab900d8 for instance with vm_state building and task_state spawning. [ 844.279184] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.279184] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 844.279184] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16740425-c249-48cd-ba14-8fb98e3fe404 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.290621] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 844.290621] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525a29a5-15ae-b1a4-0be2-6f87a4f5e676" [ 844.290621] env[61970]: _type = "Task" [ 844.290621] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.296670] env[61970]: DEBUG nova.network.neutron [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Successfully updated port: 6875c46d-b7c5-4912-a1f0-932b3ab900d8 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 844.301476] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525a29a5-15ae-b1a4-0be2-6f87a4f5e676, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.366022] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1fcf2dc9-134f-42bf-b11d-3b07d2005261 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "5d4d97e1-5812-4884-a1ea-6eb0cee591c0" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.366022] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1fcf2dc9-134f-42bf-b11d-3b07d2005261 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "5d4d97e1-5812-4884-a1ea-6eb0cee591c0" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.366022] env[61970]: DEBUG nova.compute.manager [None req-1fcf2dc9-134f-42bf-b11d-3b07d2005261 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 844.366022] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53137784-2f90-4fe9-849d-e65701b53779 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.374114] env[61970]: DEBUG nova.compute.manager [None req-1fcf2dc9-134f-42bf-b11d-3b07d2005261 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61970) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 844.374114] env[61970]: DEBUG nova.objects.instance [None req-1fcf2dc9-134f-42bf-b11d-3b07d2005261 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lazy-loading 'flavor' on Instance uuid 5d4d97e1-5812-4884-a1ea-6eb0cee591c0 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 844.503979] env[61970]: DEBUG nova.compute.utils [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 844.507497] env[61970]: DEBUG nova.compute.manager [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 844.508045] env[61970]: DEBUG nova.network.neutron [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 844.557713] env[61970]: DEBUG nova.policy [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc66d17113d447e98d8084c1b03eddb3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c0429db7acc24d2eabeaad9efa9e66ef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 844.601662] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]522b7d17-7c4a-5461-155e-7efd10d652b5, 'name': SearchDatastore_Task, 'duration_secs': 0.200642} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.603863] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.604269] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.604618] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.617415] env[61970]: DEBUG nova.network.neutron [req-dcf8d03f-d7be-4e83-a42d-6592e2bd1d28 req-5d3c571e-15cf-4db9-9c60-9be269eb5774 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Updated VIF entry in instance network info cache for port 64a3e684-f5d3-4dd2-b8c9-dc46565f71bd. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 844.617415] env[61970]: DEBUG nova.network.neutron [req-dcf8d03f-d7be-4e83-a42d-6592e2bd1d28 req-5d3c571e-15cf-4db9-9c60-9be269eb5774 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Updating instance_info_cache with network_info: [{"id": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "address": "fa:16:3e:14:8c:ec", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64a3e684-f5", "ovs_interfaceid": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.803179] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525a29a5-15ae-b1a4-0be2-6f87a4f5e676, 'name': SearchDatastore_Task, 'duration_secs': 0.038302} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.803179] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31747d47-4ef0-48ec-a723-9f98c4d45cab {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.805524] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Acquiring lock "refresh_cache-b385407b-1bdd-4c53-907c-cb4c8ce16cc7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.808017] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Acquired lock "refresh_cache-b385407b-1bdd-4c53-907c-cb4c8ce16cc7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.808017] env[61970]: DEBUG nova.network.neutron [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 844.819038] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 844.819038] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52fb5242-a1cf-840a-5c3c-d475900e4b6e" [ 844.819038] env[61970]: _type = "Task" [ 844.819038] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.831701] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52fb5242-a1cf-840a-5c3c-d475900e4b6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.873161] env[61970]: DEBUG nova.network.neutron [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Successfully created port: 1dfb626c-a7e0-455e-bffd-3eabada94ae1 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 844.880329] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fcf2dc9-134f-42bf-b11d-3b07d2005261 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 844.880621] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-951ea632-0ac1-4e9b-8d23-3b02380baf69 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.888267] env[61970]: DEBUG oslo_vmware.api [None req-1fcf2dc9-134f-42bf-b11d-3b07d2005261 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 844.888267] env[61970]: value = "task-1355614" [ 844.888267] env[61970]: _type = "Task" [ 844.888267] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.897906] env[61970]: DEBUG oslo_vmware.api [None req-1fcf2dc9-134f-42bf-b11d-3b07d2005261 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355614, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.008802] env[61970]: DEBUG nova.compute.manager [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 845.119413] env[61970]: DEBUG oslo_concurrency.lockutils [req-dcf8d03f-d7be-4e83-a42d-6592e2bd1d28 req-5d3c571e-15cf-4db9-9c60-9be269eb5774 service nova] Releasing lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.254645] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a3b5ed-0e15-49c7-8d23-c0199eeca153 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.262319] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb475fc-91a3-4914-8ed7-0dac691a76cc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.292135] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6540871-379f-4b11-9b8a-477c26d9e26b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.299873] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0823baca-9a57-45ee-a167-0de1f8bb675e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.318404] env[61970]: DEBUG nova.compute.provider_tree [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.329881] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52fb5242-a1cf-840a-5c3c-d475900e4b6e, 'name': SearchDatastore_Task, 'duration_secs': 0.155535} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.330463] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.330848] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] b19cb516-b163-4bed-ba5b-139a0a18fc05/b19cb516-b163-4bed-ba5b-139a0a18fc05.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 845.331156] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.331360] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.331823] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-be011e42-f3bc-4012-b2ab-2bfd0c62d3f4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.333689] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3e4dcefd-21ea-482f-b085-a16a3991f8b1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.342674] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 845.342674] env[61970]: value = "task-1355615" [ 845.342674] env[61970]: _type = "Task" [ 845.342674] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.350733] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355615, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.356358] env[61970]: DEBUG nova.network.neutron [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 845.397558] env[61970]: DEBUG oslo_vmware.api [None req-1fcf2dc9-134f-42bf-b11d-3b07d2005261 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355614, 'name': PowerOffVM_Task, 'duration_secs': 0.433561} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.397881] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fcf2dc9-134f-42bf-b11d-3b07d2005261 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 845.398105] env[61970]: DEBUG nova.compute.manager [None req-1fcf2dc9-134f-42bf-b11d-3b07d2005261 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 845.399291] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f190e515-8e52-40cb-90c5-95846332d542 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.514076] env[61970]: DEBUG nova.network.neutron [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Updating instance_info_cache with network_info: [{"id": "6875c46d-b7c5-4912-a1f0-932b3ab900d8", "address": "fa:16:3e:00:b6:23", "network": {"id": "01909a13-6915-4607-b08d-9d5aac1ab78f", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1225281584-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2c5df141b77042c28368dc4474f7d917", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "163e60bd-32d6-41c5-95e6-2eb10c5c9245", "external-id": "nsx-vlan-transportzone-716", "segmentation_id": 716, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6875c46d-b7", "ovs_interfaceid": "6875c46d-b7c5-4912-a1f0-932b3ab900d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.517574] env[61970]: INFO nova.virt.block_device [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Booting with volume 4020284d-a273-459a-bb8c-84753ab98b1e at /dev/sda [ 845.560040] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1cbefaf7-bba4-4a5c-ad33-e873a68c805e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.569427] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3bca894-7762-4003-bc62-506fbfc880d3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.595119] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2eb3b091-c1be-46bd-a3de-c309558a0638 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.602787] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef80a743-a790-4621-a693-8310f9cd85d0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.614879] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.615079] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 845.615772] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f54ccbe-ecbb-46c2-ac12-efada980ee50 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.620741] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 845.620741] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52216e40-1e0f-65a5-efd1-d33d3f159431" [ 845.620741] env[61970]: _type = "Task" [ 845.620741] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.630367] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49718224-e539-4d3d-aa42-b8c6f5d2e556 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.635410] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52216e40-1e0f-65a5-efd1-d33d3f159431, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.639535] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3299eb5-2209-45e5-820c-6e4fa9f202c0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.652899] env[61970]: DEBUG nova.virt.block_device [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Updating existing volume attachment record: 6a34de95-2724-44f0-b3ce-36ea2289a250 {{(pid=61970) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 845.822045] env[61970]: DEBUG nova.scheduler.client.report [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.854453] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355615, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.911929] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1fcf2dc9-134f-42bf-b11d-3b07d2005261 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "5d4d97e1-5812-4884-a1ea-6eb0cee591c0" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.548s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.017560] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Releasing lock "refresh_cache-b385407b-1bdd-4c53-907c-cb4c8ce16cc7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.017889] env[61970]: DEBUG nova.compute.manager [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Instance network_info: |[{"id": "6875c46d-b7c5-4912-a1f0-932b3ab900d8", "address": "fa:16:3e:00:b6:23", "network": {"id": "01909a13-6915-4607-b08d-9d5aac1ab78f", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1225281584-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2c5df141b77042c28368dc4474f7d917", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "163e60bd-32d6-41c5-95e6-2eb10c5c9245", "external-id": "nsx-vlan-transportzone-716", "segmentation_id": 716, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6875c46d-b7", "ovs_interfaceid": "6875c46d-b7c5-4912-a1f0-932b3ab900d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 846.018353] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:00:b6:23', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '163e60bd-32d6-41c5-95e6-2eb10c5c9245', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6875c46d-b7c5-4912-a1f0-932b3ab900d8', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 846.026703] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Creating folder: Project (2c5df141b77042c28368dc4474f7d917). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 846.026979] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1553982c-bd8a-4f8f-98e2-67ade7816a4a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.039554] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Created folder: Project (2c5df141b77042c28368dc4474f7d917) in parent group-v288740. [ 846.039745] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Creating folder: Instances. Parent ref: group-v288781. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 846.039977] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a9075a78-d7a1-48a7-a792-ec516de86210 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.049166] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Created folder: Instances in parent group-v288781. [ 846.049390] env[61970]: DEBUG oslo.service.loopingcall [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.049575] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 846.049768] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5c1e8326-3cdb-4b5e-9f7d-26ff98481997 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.068250] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 846.068250] env[61970]: value = "task-1355618" [ 846.068250] env[61970]: _type = "Task" [ 846.068250] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.075448] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355618, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.130837] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52216e40-1e0f-65a5-efd1-d33d3f159431, 'name': SearchDatastore_Task, 'duration_secs': 0.414497} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.131792] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bf3b29d-1bfd-41d7-bc55-f12b7d275db4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.137397] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 846.137397] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5203b00c-1a82-39dc-2439-09426ce7bc39" [ 846.137397] env[61970]: _type = "Task" [ 846.137397] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.144987] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5203b00c-1a82-39dc-2439-09426ce7bc39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.183794] env[61970]: DEBUG nova.compute.manager [req-96a3ff25-d05e-4379-8fdb-937cfe8bc22c req-b45ece9e-a3a4-4463-b874-0cf93a477b8d service nova] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Received event network-changed-6875c46d-b7c5-4912-a1f0-932b3ab900d8 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.184103] env[61970]: DEBUG nova.compute.manager [req-96a3ff25-d05e-4379-8fdb-937cfe8bc22c req-b45ece9e-a3a4-4463-b874-0cf93a477b8d service nova] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Refreshing instance network info cache due to event network-changed-6875c46d-b7c5-4912-a1f0-932b3ab900d8. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 846.184371] env[61970]: DEBUG oslo_concurrency.lockutils [req-96a3ff25-d05e-4379-8fdb-937cfe8bc22c req-b45ece9e-a3a4-4463-b874-0cf93a477b8d service nova] Acquiring lock "refresh_cache-b385407b-1bdd-4c53-907c-cb4c8ce16cc7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.184568] env[61970]: DEBUG oslo_concurrency.lockutils [req-96a3ff25-d05e-4379-8fdb-937cfe8bc22c req-b45ece9e-a3a4-4463-b874-0cf93a477b8d service nova] Acquired lock "refresh_cache-b385407b-1bdd-4c53-907c-cb4c8ce16cc7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.184745] env[61970]: DEBUG nova.network.neutron [req-96a3ff25-d05e-4379-8fdb-937cfe8bc22c req-b45ece9e-a3a4-4463-b874-0cf93a477b8d service nova] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Refreshing network info cache for port 6875c46d-b7c5-4912-a1f0-932b3ab900d8 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 846.327076] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.335s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.327654] env[61970]: DEBUG nova.compute.manager [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 846.330746] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.701s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.357274] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355615, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.368908] env[61970]: DEBUG nova.compute.manager [req-45e90530-37e5-4640-97c9-723ace499d50 req-6a7d66bf-a272-4cbe-b9fb-3600b90ac364 service nova] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Received event network-vif-plugged-1dfb626c-a7e0-455e-bffd-3eabada94ae1 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.369189] env[61970]: DEBUG oslo_concurrency.lockutils [req-45e90530-37e5-4640-97c9-723ace499d50 req-6a7d66bf-a272-4cbe-b9fb-3600b90ac364 service nova] Acquiring lock "6838c6a4-8241-4b6f-ad58-3acfd7afdbf4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.369467] env[61970]: DEBUG oslo_concurrency.lockutils [req-45e90530-37e5-4640-97c9-723ace499d50 req-6a7d66bf-a272-4cbe-b9fb-3600b90ac364 service nova] Lock "6838c6a4-8241-4b6f-ad58-3acfd7afdbf4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.369705] env[61970]: DEBUG oslo_concurrency.lockutils [req-45e90530-37e5-4640-97c9-723ace499d50 req-6a7d66bf-a272-4cbe-b9fb-3600b90ac364 service nova] Lock "6838c6a4-8241-4b6f-ad58-3acfd7afdbf4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.369849] env[61970]: DEBUG nova.compute.manager [req-45e90530-37e5-4640-97c9-723ace499d50 req-6a7d66bf-a272-4cbe-b9fb-3600b90ac364 service nova] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] No waiting events found dispatching network-vif-plugged-1dfb626c-a7e0-455e-bffd-3eabada94ae1 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 846.370055] env[61970]: WARNING nova.compute.manager [req-45e90530-37e5-4640-97c9-723ace499d50 req-6a7d66bf-a272-4cbe-b9fb-3600b90ac364 service nova] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Received unexpected event network-vif-plugged-1dfb626c-a7e0-455e-bffd-3eabada94ae1 for instance with vm_state building and task_state block_device_mapping. [ 846.471193] env[61970]: DEBUG nova.network.neutron [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Successfully updated port: 1dfb626c-a7e0-455e-bffd-3eabada94ae1 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 846.577826] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355618, 'name': CreateVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.649498] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5203b00c-1a82-39dc-2439-09426ce7bc39, 'name': SearchDatastore_Task, 'duration_secs': 0.020039} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.649760] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.650015] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] b1cc1cc2-15d6-459d-9529-e592ddb225ac/b1cc1cc2-15d6-459d-9529-e592ddb225ac.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 846.650287] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-53992589-6f49-4909-bdfa-de3e57fbc922 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.658016] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 846.658016] env[61970]: value = "task-1355619" [ 846.658016] env[61970]: _type = "Task" [ 846.658016] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.665935] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355619, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.835962] env[61970]: DEBUG nova.compute.utils [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 846.840371] env[61970]: DEBUG nova.compute.manager [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 846.840544] env[61970]: DEBUG nova.network.neutron [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 846.843169] env[61970]: DEBUG nova.objects.instance [None req-25171177-b508-4c7d-af68-a3f2c99295fb tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lazy-loading 'flavor' on Instance uuid 5d4d97e1-5812-4884-a1ea-6eb0cee591c0 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 846.854580] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355615, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.19795} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.854817] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] b19cb516-b163-4bed-ba5b-139a0a18fc05/b19cb516-b163-4bed-ba5b-139a0a18fc05.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 846.855067] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.855332] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e54f20d5-2b73-48db-9bdf-3782246644a7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.864808] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 846.864808] env[61970]: value = "task-1355620" [ 846.864808] env[61970]: _type = "Task" [ 846.864808] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.877638] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355620, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.886063] env[61970]: DEBUG nova.policy [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64082edf5f2e4621b5f3586e1792e42a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9235eeeb6fa4ccf959ffbea456f3694', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 846.926266] env[61970]: DEBUG nova.network.neutron [req-96a3ff25-d05e-4379-8fdb-937cfe8bc22c req-b45ece9e-a3a4-4463-b874-0cf93a477b8d service nova] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Updated VIF entry in instance network info cache for port 6875c46d-b7c5-4912-a1f0-932b3ab900d8. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 846.926620] env[61970]: DEBUG nova.network.neutron [req-96a3ff25-d05e-4379-8fdb-937cfe8bc22c req-b45ece9e-a3a4-4463-b874-0cf93a477b8d service nova] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Updating instance_info_cache with network_info: [{"id": "6875c46d-b7c5-4912-a1f0-932b3ab900d8", "address": "fa:16:3e:00:b6:23", "network": {"id": "01909a13-6915-4607-b08d-9d5aac1ab78f", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1225281584-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2c5df141b77042c28368dc4474f7d917", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "163e60bd-32d6-41c5-95e6-2eb10c5c9245", "external-id": "nsx-vlan-transportzone-716", "segmentation_id": 716, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6875c46d-b7", "ovs_interfaceid": "6875c46d-b7c5-4912-a1f0-932b3ab900d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.974342] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Acquiring lock "refresh_cache-6838c6a4-8241-4b6f-ad58-3acfd7afdbf4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.974506] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Acquired lock "refresh_cache-6838c6a4-8241-4b6f-ad58-3acfd7afdbf4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.974655] env[61970]: DEBUG nova.network.neutron [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 847.076138] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4be09a-2d54-4801-900a-5fa70f0e1c6f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.081713] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355618, 'name': CreateVM_Task, 'duration_secs': 0.572536} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.082282] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 847.082969] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.083146] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.083466] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 847.083774] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98f7fa7e-3822-4128-9940-836610f3e4d9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.087992] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08af064-b921-4904-8484-b0c7c20aab74 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.093156] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Waiting for the task: (returnval){ [ 847.093156] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523f21a1-3310-251e-a636-2085477d96f2" [ 847.093156] env[61970]: _type = "Task" [ 847.093156] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.128133] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d508b5-de41-48f1-a59d-d49703cd92eb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.136993] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523f21a1-3310-251e-a636-2085477d96f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.139308] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42a445a-6286-47dd-bb1c-4feb2f4d911c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.153668] env[61970]: DEBUG nova.compute.provider_tree [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.155669] env[61970]: DEBUG nova.network.neutron [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Successfully created port: 201e52ca-273b-4df3-b32a-d9dba51cb5c7 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 847.169332] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355619, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.346020] env[61970]: DEBUG nova.compute.manager [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 847.350899] env[61970]: DEBUG oslo_concurrency.lockutils [None req-25171177-b508-4c7d-af68-a3f2c99295fb tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "refresh_cache-5d4d97e1-5812-4884-a1ea-6eb0cee591c0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.351218] env[61970]: DEBUG oslo_concurrency.lockutils [None req-25171177-b508-4c7d-af68-a3f2c99295fb tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquired lock "refresh_cache-5d4d97e1-5812-4884-a1ea-6eb0cee591c0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.351504] env[61970]: DEBUG nova.network.neutron [None req-25171177-b508-4c7d-af68-a3f2c99295fb tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 847.351771] env[61970]: DEBUG nova.objects.instance [None req-25171177-b508-4c7d-af68-a3f2c99295fb tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lazy-loading 'info_cache' on Instance uuid 5d4d97e1-5812-4884-a1ea-6eb0cee591c0 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 847.378690] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355620, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.310417} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.378880] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 847.379681] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f87bf0d-1610-4bc6-b2a2-6d418234e1b0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.401355] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] b19cb516-b163-4bed-ba5b-139a0a18fc05/b19cb516-b163-4bed-ba5b-139a0a18fc05.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.402533] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b1959cbe-21d6-4e7c-9426-837bb57bcd43 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.422265] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 847.422265] env[61970]: value = "task-1355621" [ 847.422265] env[61970]: _type = "Task" [ 847.422265] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.430307] env[61970]: DEBUG oslo_concurrency.lockutils [req-96a3ff25-d05e-4379-8fdb-937cfe8bc22c req-b45ece9e-a3a4-4463-b874-0cf93a477b8d service nova] Releasing lock "refresh_cache-b385407b-1bdd-4c53-907c-cb4c8ce16cc7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.430685] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355621, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.509721] env[61970]: DEBUG nova.network.neutron [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 847.607173] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523f21a1-3310-251e-a636-2085477d96f2, 'name': SearchDatastore_Task, 'duration_secs': 0.04267} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.607560] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.607812] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 847.608124] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.608301] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.608504] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 847.608841] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0727b1d2-a7d4-4d1d-9679-4193a3ab6176 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.623732] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 847.623986] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 847.624782] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4824df28-c39d-4cb9-821d-23333832d549 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.633336] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Waiting for the task: (returnval){ [ 847.633336] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5227f75e-11b9-51ad-e429-ed54b0f858ef" [ 847.633336] env[61970]: _type = "Task" [ 847.633336] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.644574] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5227f75e-11b9-51ad-e429-ed54b0f858ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.662425] env[61970]: DEBUG nova.scheduler.client.report [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 847.677319] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355619, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.752019] env[61970]: DEBUG nova.compute.manager [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 847.752019] env[61970]: DEBUG nova.virt.hardware [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 847.752019] env[61970]: DEBUG nova.virt.hardware [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 847.752019] env[61970]: DEBUG nova.virt.hardware [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.752321] env[61970]: DEBUG nova.virt.hardware [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 847.752321] env[61970]: DEBUG nova.virt.hardware [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.752321] env[61970]: DEBUG nova.virt.hardware [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 847.752321] env[61970]: DEBUG nova.virt.hardware [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 847.752321] env[61970]: DEBUG nova.virt.hardware [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 847.753556] env[61970]: DEBUG nova.virt.hardware [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 847.753992] env[61970]: DEBUG nova.virt.hardware [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 847.754370] env[61970]: DEBUG nova.virt.hardware [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.755453] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63aa0e8b-11ac-4613-8ec0-fa1833e1a25e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.759780] env[61970]: DEBUG nova.network.neutron [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Updating instance_info_cache with network_info: [{"id": "1dfb626c-a7e0-455e-bffd-3eabada94ae1", "address": "fa:16:3e:b9:26:72", "network": {"id": "f08966ed-3f5b-4be7-bcb0-a0fe277e5367", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1451995592-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0429db7acc24d2eabeaad9efa9e66ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "650f7968-4522-4ba5-8304-1b9949951ed7", "external-id": "nsx-vlan-transportzone-568", "segmentation_id": 568, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1dfb626c-a7", "ovs_interfaceid": "1dfb626c-a7e0-455e-bffd-3eabada94ae1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.767104] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53625be7-b864-45bf-ae54-cf24828e011c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.857740] env[61970]: DEBUG nova.objects.base [None req-25171177-b508-4c7d-af68-a3f2c99295fb tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Object Instance<5d4d97e1-5812-4884-a1ea-6eb0cee591c0> lazy-loaded attributes: flavor,info_cache {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 847.932457] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355621, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.143885] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5227f75e-11b9-51ad-e429-ed54b0f858ef, 'name': SearchDatastore_Task, 'duration_secs': 0.054618} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.144179] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e58781e-9c66-46d3-b8dd-2d13071df584 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.149259] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Waiting for the task: (returnval){ [ 848.149259] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52dc0902-4f61-f503-df98-40c6f162732f" [ 848.149259] env[61970]: _type = "Task" [ 848.149259] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.156469] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52dc0902-4f61-f503-df98-40c6f162732f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.170977] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.840s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.171675] env[61970]: ERROR nova.compute.manager [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1187b71f-0e90-465d-b15a-d9eae922f9de, please check neutron logs for more information. [ 848.171675] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Traceback (most recent call last): [ 848.171675] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 848.171675] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] self.driver.spawn(context, instance, image_meta, [ 848.171675] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 848.171675] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 848.171675] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 848.171675] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] vm_ref = self.build_virtual_machine(instance, [ 848.171675] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 848.171675] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] vif_infos = vmwarevif.get_vif_info(self._session, [ 848.171675] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 848.172075] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] for vif in network_info: [ 848.172075] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 848.172075] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] return self._sync_wrapper(fn, *args, **kwargs) [ 848.172075] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 848.172075] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] self.wait() [ 848.172075] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 848.172075] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] self[:] = self._gt.wait() [ 848.172075] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 848.172075] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] return self._exit_event.wait() [ 848.172075] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 848.172075] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] current.throw(*self._exc) [ 848.172075] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 848.172075] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] result = function(*args, **kwargs) [ 848.172469] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 848.172469] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] return func(*args, **kwargs) [ 848.172469] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 848.172469] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] raise e [ 848.172469] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 848.172469] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] nwinfo = self.network_api.allocate_for_instance( [ 848.172469] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 848.172469] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] created_port_ids = self._update_ports_for_instance( [ 848.172469] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 848.172469] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] with excutils.save_and_reraise_exception(): [ 848.172469] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 848.172469] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] self.force_reraise() [ 848.172469] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 848.172849] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] raise self.value [ 848.172849] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 848.172849] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] updated_port = self._update_port( [ 848.172849] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 848.172849] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] _ensure_no_port_binding_failure(port) [ 848.172849] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 848.172849] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] raise exception.PortBindingFailed(port_id=port['id']) [ 848.172849] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] nova.exception.PortBindingFailed: Binding failed for port 1187b71f-0e90-465d-b15a-d9eae922f9de, please check neutron logs for more information. [ 848.172849] env[61970]: ERROR nova.compute.manager [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] [ 848.172849] env[61970]: DEBUG nova.compute.utils [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Binding failed for port 1187b71f-0e90-465d-b15a-d9eae922f9de, please check neutron logs for more information. {{(pid=61970) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 848.176908] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.443s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.177117] env[61970]: DEBUG nova.objects.instance [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61970) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 848.179909] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355619, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.114793} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.180567] env[61970]: DEBUG nova.compute.manager [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Build of instance c5c253b6-ebab-4fea-bcc4-f3bb3960070b was re-scheduled: Binding failed for port 1187b71f-0e90-465d-b15a-d9eae922f9de, please check neutron logs for more information. {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 848.180731] env[61970]: DEBUG nova.compute.manager [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Unplugging VIFs for instance {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 848.180862] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Acquiring lock "refresh_cache-c5c253b6-ebab-4fea-bcc4-f3bb3960070b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.181032] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Acquired lock "refresh_cache-c5c253b6-ebab-4fea-bcc4-f3bb3960070b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.181201] env[61970]: DEBUG nova.network.neutron [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 848.182358] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] b1cc1cc2-15d6-459d-9529-e592ddb225ac/b1cc1cc2-15d6-459d-9529-e592ddb225ac.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 848.182595] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 848.182986] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6c8cf3f0-88e2-4ec3-a504-2a16b9a213f4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.190830] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 848.190830] env[61970]: value = "task-1355622" [ 848.190830] env[61970]: _type = "Task" [ 848.190830] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.200831] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355622, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.262345] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Releasing lock "refresh_cache-6838c6a4-8241-4b6f-ad58-3acfd7afdbf4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.262674] env[61970]: DEBUG nova.compute.manager [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Instance network_info: |[{"id": "1dfb626c-a7e0-455e-bffd-3eabada94ae1", "address": "fa:16:3e:b9:26:72", "network": {"id": "f08966ed-3f5b-4be7-bcb0-a0fe277e5367", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1451995592-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0429db7acc24d2eabeaad9efa9e66ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "650f7968-4522-4ba5-8304-1b9949951ed7", "external-id": "nsx-vlan-transportzone-568", "segmentation_id": 568, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1dfb626c-a7", "ovs_interfaceid": "1dfb626c-a7e0-455e-bffd-3eabada94ae1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 848.263104] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:26:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '650f7968-4522-4ba5-8304-1b9949951ed7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1dfb626c-a7e0-455e-bffd-3eabada94ae1', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 848.270533] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Creating folder: Project (c0429db7acc24d2eabeaad9efa9e66ef). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 848.270853] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-71a30519-7040-43c0-a8b8-d51832b6af43 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.284726] env[61970]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 848.284880] env[61970]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61970) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 848.285281] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Folder already exists: Project (c0429db7acc24d2eabeaad9efa9e66ef). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 848.285480] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Creating folder: Instances. Parent ref: group-v288761. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 848.285741] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d97fa901-d566-491e-9695-f0a0aa31da54 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.295570] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Created folder: Instances in parent group-v288761. [ 848.295794] env[61970]: DEBUG oslo.service.loopingcall [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 848.295978] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 848.296195] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9a4680d8-77fb-4ebd-8b55-9e6e2b57e89f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.315920] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 848.315920] env[61970]: value = "task-1355625" [ 848.315920] env[61970]: _type = "Task" [ 848.315920] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.323641] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355625, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.358573] env[61970]: DEBUG nova.compute.manager [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 848.388924] env[61970]: DEBUG nova.virt.hardware [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 848.389252] env[61970]: DEBUG nova.virt.hardware [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 848.389436] env[61970]: DEBUG nova.virt.hardware [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 848.389908] env[61970]: DEBUG nova.virt.hardware [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 848.389908] env[61970]: DEBUG nova.virt.hardware [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 848.390015] env[61970]: DEBUG nova.virt.hardware [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 848.391025] env[61970]: DEBUG nova.virt.hardware [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 848.391025] env[61970]: DEBUG nova.virt.hardware [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 848.391025] env[61970]: DEBUG nova.virt.hardware [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 848.391025] env[61970]: DEBUG nova.virt.hardware [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 848.391025] env[61970]: DEBUG nova.virt.hardware [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 848.391901] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca3c1fcf-9940-4e06-8d54-346b6d852e0e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.396763] env[61970]: DEBUG nova.compute.manager [req-b10cb5c2-f366-4d0d-a850-8eb46392b77b req-b0843409-78a3-4ff7-843b-9a97cb043127 service nova] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Received event network-changed-1dfb626c-a7e0-455e-bffd-3eabada94ae1 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.396955] env[61970]: DEBUG nova.compute.manager [req-b10cb5c2-f366-4d0d-a850-8eb46392b77b req-b0843409-78a3-4ff7-843b-9a97cb043127 service nova] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Refreshing instance network info cache due to event network-changed-1dfb626c-a7e0-455e-bffd-3eabada94ae1. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 848.397186] env[61970]: DEBUG oslo_concurrency.lockutils [req-b10cb5c2-f366-4d0d-a850-8eb46392b77b req-b0843409-78a3-4ff7-843b-9a97cb043127 service nova] Acquiring lock "refresh_cache-6838c6a4-8241-4b6f-ad58-3acfd7afdbf4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.397327] env[61970]: DEBUG oslo_concurrency.lockutils [req-b10cb5c2-f366-4d0d-a850-8eb46392b77b req-b0843409-78a3-4ff7-843b-9a97cb043127 service nova] Acquired lock "refresh_cache-6838c6a4-8241-4b6f-ad58-3acfd7afdbf4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.397487] env[61970]: DEBUG nova.network.neutron [req-b10cb5c2-f366-4d0d-a850-8eb46392b77b req-b0843409-78a3-4ff7-843b-9a97cb043127 service nova] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Refreshing network info cache for port 1dfb626c-a7e0-455e-bffd-3eabada94ae1 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 848.407227] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c83bce-4ff9-4fda-a946-2c234ee832cb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.431996] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355621, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.583530] env[61970]: DEBUG nova.network.neutron [None req-25171177-b508-4c7d-af68-a3f2c99295fb tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Updating instance_info_cache with network_info: [{"id": "1080e2f3-c9a3-48bc-bd65-b5f231bacdc7", "address": "fa:16:3e:55:7f:ba", "network": {"id": "700300d5-dfe3-4ff7-86ed-43a4f2c15818", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1754634816-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e91895bd10c74f8d9a101274d87cdc71", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ba56987-9dc3-4c76-a4e2-942b05355bdb", "external-id": "nsx-vlan-transportzone-698", "segmentation_id": 698, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1080e2f3-c9", "ovs_interfaceid": "1080e2f3-c9a3-48bc-bd65-b5f231bacdc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.591183] env[61970]: DEBUG nova.compute.manager [req-ba1efeae-160b-49d5-b82e-3d398a920459 req-4f9d48be-ffbf-4e21-9d96-52f250f2bb85 service nova] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Received event network-vif-plugged-201e52ca-273b-4df3-b32a-d9dba51cb5c7 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.591343] env[61970]: DEBUG oslo_concurrency.lockutils [req-ba1efeae-160b-49d5-b82e-3d398a920459 req-4f9d48be-ffbf-4e21-9d96-52f250f2bb85 service nova] Acquiring lock "d09dbde2-5daa-4ae9-9fd5-bfbb03fba136-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.591536] env[61970]: DEBUG oslo_concurrency.lockutils [req-ba1efeae-160b-49d5-b82e-3d398a920459 req-4f9d48be-ffbf-4e21-9d96-52f250f2bb85 service nova] Lock "d09dbde2-5daa-4ae9-9fd5-bfbb03fba136-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.591663] env[61970]: DEBUG oslo_concurrency.lockutils [req-ba1efeae-160b-49d5-b82e-3d398a920459 req-4f9d48be-ffbf-4e21-9d96-52f250f2bb85 service nova] Lock "d09dbde2-5daa-4ae9-9fd5-bfbb03fba136-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.591833] env[61970]: DEBUG nova.compute.manager [req-ba1efeae-160b-49d5-b82e-3d398a920459 req-4f9d48be-ffbf-4e21-9d96-52f250f2bb85 service nova] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] No waiting events found dispatching network-vif-plugged-201e52ca-273b-4df3-b32a-d9dba51cb5c7 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 848.592185] env[61970]: WARNING nova.compute.manager [req-ba1efeae-160b-49d5-b82e-3d398a920459 req-4f9d48be-ffbf-4e21-9d96-52f250f2bb85 service nova] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Received unexpected event network-vif-plugged-201e52ca-273b-4df3-b32a-d9dba51cb5c7 for instance with vm_state building and task_state spawning. [ 848.660081] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52dc0902-4f61-f503-df98-40c6f162732f, 'name': SearchDatastore_Task, 'duration_secs': 0.018711} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.660592] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.660842] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] b385407b-1bdd-4c53-907c-cb4c8ce16cc7/b385407b-1bdd-4c53-907c-cb4c8ce16cc7.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 848.661106] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-238f3a3f-548b-4392-a999-c0e8f506d907 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.667919] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Waiting for the task: (returnval){ [ 848.667919] env[61970]: value = "task-1355626" [ 848.667919] env[61970]: _type = "Task" [ 848.667919] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.675516] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': task-1355626, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.680163] env[61970]: DEBUG nova.network.neutron [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Successfully updated port: 201e52ca-273b-4df3-b32a-d9dba51cb5c7 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 848.701343] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355622, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065522} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.702156] env[61970]: DEBUG nova.network.neutron [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 848.703828] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 848.704659] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb488f1-7be2-4e87-8025-bdb69ce191b5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.728721] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] b1cc1cc2-15d6-459d-9529-e592ddb225ac/b1cc1cc2-15d6-459d-9529-e592ddb225ac.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.729367] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac49c250-cc41-4912-ad41-309886623701 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.749071] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 848.749071] env[61970]: value = "task-1355627" [ 848.749071] env[61970]: _type = "Task" [ 848.749071] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.756679] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355627, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.811809] env[61970]: DEBUG nova.network.neutron [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.826158] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355625, 'name': CreateVM_Task} progress is 25%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.937586] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355621, 'name': ReconfigVM_Task, 'duration_secs': 1.468263} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.937955] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Reconfigured VM instance instance-00000040 to attach disk [datastore1] b19cb516-b163-4bed-ba5b-139a0a18fc05/b19cb516-b163-4bed-ba5b-139a0a18fc05.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.938839] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-59e5074a-4aee-4a4a-981c-d2197e1703cf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.945929] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 848.945929] env[61970]: value = "task-1355628" [ 848.945929] env[61970]: _type = "Task" [ 848.945929] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.963275] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355628, 'name': Rename_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.086228] env[61970]: DEBUG oslo_concurrency.lockutils [None req-25171177-b508-4c7d-af68-a3f2c99295fb tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Releasing lock "refresh_cache-5d4d97e1-5812-4884-a1ea-6eb0cee591c0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.179723] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': task-1355626, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.182988] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "refresh_cache-d09dbde2-5daa-4ae9-9fd5-bfbb03fba136" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.183100] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "refresh_cache-d09dbde2-5daa-4ae9-9fd5-bfbb03fba136" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.183496] env[61970]: DEBUG nova.network.neutron [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 849.191740] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5b862e26-7907-47da-a7b1-0ad203c8dd47 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.191740] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.255s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.191740] env[61970]: DEBUG nova.objects.instance [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lazy-loading 'resources' on Instance uuid a78b63a5-6bb8-4271-90d8-1e86fb29db4f {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.221610] env[61970]: DEBUG nova.network.neutron [req-b10cb5c2-f366-4d0d-a850-8eb46392b77b req-b0843409-78a3-4ff7-843b-9a97cb043127 service nova] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Updated VIF entry in instance network info cache for port 1dfb626c-a7e0-455e-bffd-3eabada94ae1. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 849.221730] env[61970]: DEBUG nova.network.neutron [req-b10cb5c2-f366-4d0d-a850-8eb46392b77b req-b0843409-78a3-4ff7-843b-9a97cb043127 service nova] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Updating instance_info_cache with network_info: [{"id": "1dfb626c-a7e0-455e-bffd-3eabada94ae1", "address": "fa:16:3e:b9:26:72", "network": {"id": "f08966ed-3f5b-4be7-bcb0-a0fe277e5367", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1451995592-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0429db7acc24d2eabeaad9efa9e66ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "650f7968-4522-4ba5-8304-1b9949951ed7", "external-id": "nsx-vlan-transportzone-568", "segmentation_id": 568, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1dfb626c-a7", "ovs_interfaceid": "1dfb626c-a7e0-455e-bffd-3eabada94ae1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.260681] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355627, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.315712] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Releasing lock "refresh_cache-c5c253b6-ebab-4fea-bcc4-f3bb3960070b" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.315964] env[61970]: DEBUG nova.compute.manager [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61970) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 849.316215] env[61970]: DEBUG nova.compute.manager [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 849.316661] env[61970]: DEBUG nova.network.neutron [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 849.328526] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355625, 'name': CreateVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.340657] env[61970]: DEBUG nova.network.neutron [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 849.459604] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355628, 'name': Rename_Task, 'duration_secs': 0.288698} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.459872] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 849.460145] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-79b76eff-fc73-439f-ab19-ee6ac46d9f6e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.467063] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 849.467063] env[61970]: value = "task-1355629" [ 849.467063] env[61970]: _type = "Task" [ 849.467063] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.475595] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355629, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.590350] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-25171177-b508-4c7d-af68-a3f2c99295fb tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 849.590668] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cc74dd64-00f4-4184-a1f6-b977776e729d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.598206] env[61970]: DEBUG oslo_vmware.api [None req-25171177-b508-4c7d-af68-a3f2c99295fb tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 849.598206] env[61970]: value = "task-1355630" [ 849.598206] env[61970]: _type = "Task" [ 849.598206] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.606441] env[61970]: DEBUG oslo_vmware.api [None req-25171177-b508-4c7d-af68-a3f2c99295fb tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355630, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.678030] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': task-1355626, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.604974} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.678327] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] b385407b-1bdd-4c53-907c-cb4c8ce16cc7/b385407b-1bdd-4c53-907c-cb4c8ce16cc7.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 849.678576] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 849.678849] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-439627fb-a796-41ba-8daa-300b25c98a76 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.685119] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Waiting for the task: (returnval){ [ 849.685119] env[61970]: value = "task-1355631" [ 849.685119] env[61970]: _type = "Task" [ 849.685119] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.696895] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': task-1355631, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.724903] env[61970]: DEBUG oslo_concurrency.lockutils [req-b10cb5c2-f366-4d0d-a850-8eb46392b77b req-b0843409-78a3-4ff7-843b-9a97cb043127 service nova] Releasing lock "refresh_cache-6838c6a4-8241-4b6f-ad58-3acfd7afdbf4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.731528] env[61970]: DEBUG nova.network.neutron [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 849.761052] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355627, 'name': ReconfigVM_Task, 'duration_secs': 0.588269} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.761338] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Reconfigured VM instance instance-00000041 to attach disk [datastore1] b1cc1cc2-15d6-459d-9529-e592ddb225ac/b1cc1cc2-15d6-459d-9529-e592ddb225ac.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 849.761941] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c7706fc0-3ee7-4af7-920a-8045037dcd74 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.767966] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 849.767966] env[61970]: value = "task-1355632" [ 849.767966] env[61970]: _type = "Task" [ 849.767966] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.778829] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355632, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.831834] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355625, 'name': CreateVM_Task, 'duration_secs': 1.17075} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.832043] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 849.832685] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': True, 'attachment_id': '6a34de95-2724-44f0-b3ce-36ea2289a250', 'boot_index': 0, 'device_type': None, 'mount_device': '/dev/sda', 'disk_bus': None, 'guest_format': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288764', 'volume_id': '4020284d-a273-459a-bb8c-84753ab98b1e', 'name': 'volume-4020284d-a273-459a-bb8c-84753ab98b1e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6838c6a4-8241-4b6f-ad58-3acfd7afdbf4', 'attached_at': '', 'detached_at': '', 'volume_id': '4020284d-a273-459a-bb8c-84753ab98b1e', 'serial': '4020284d-a273-459a-bb8c-84753ab98b1e'}, 'volume_type': None}], 'swap': None} {{(pid=61970) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 849.832891] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Root volume attach. Driver type: vmdk {{(pid=61970) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 849.834036] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796fbdbf-9fd6-4142-bada-d94884f17543 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.843159] env[61970]: DEBUG nova.network.neutron [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.844691] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af91aa0-38c6-4c40-bfc7-731b7f16d81e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.855469] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3172a584-3064-41c4-897b-6d3ee7ae7d94 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.864728] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-ddb5d949-046e-4311-b6c8-f4438437b929 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.872694] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Waiting for the task: (returnval){ [ 849.872694] env[61970]: value = "task-1355633" [ 849.872694] env[61970]: _type = "Task" [ 849.872694] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.884528] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355633, 'name': RelocateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.908108] env[61970]: DEBUG nova.network.neutron [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Updating instance_info_cache with network_info: [{"id": "201e52ca-273b-4df3-b32a-d9dba51cb5c7", "address": "fa:16:3e:d1:66:5f", "network": {"id": "e02bfa6c-8f0d-4413-8dcf-e644c07792d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1410864732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9235eeeb6fa4ccf959ffbea456f3694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap201e52ca-27", "ovs_interfaceid": "201e52ca-273b-4df3-b32a-d9dba51cb5c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.969801] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-966cb096-02cf-47fc-bf4a-0382c32fed66 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.988324] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e91ccb-949b-4507-b13f-73940d8f978c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.993352] env[61970]: DEBUG oslo_vmware.api [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355629, 'name': PowerOnVM_Task, 'duration_secs': 0.458222} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.993715] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 849.994125] env[61970]: INFO nova.compute.manager [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Took 11.19 seconds to spawn the instance on the hypervisor. [ 849.994287] env[61970]: DEBUG nova.compute.manager [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.996208] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9cbfd81-f581-4012-b77f-d732474b27c2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.032788] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4635e9d5-1221-417d-a72d-3e84c3d76bc9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.043723] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c0f78d-86c4-4ddb-b25b-5a7741552953 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.058696] env[61970]: DEBUG nova.compute.provider_tree [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 850.111548] env[61970]: DEBUG oslo_vmware.api [None req-25171177-b508-4c7d-af68-a3f2c99295fb tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355630, 'name': PowerOnVM_Task, 'duration_secs': 0.406541} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.111971] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-25171177-b508-4c7d-af68-a3f2c99295fb tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 850.112289] env[61970]: DEBUG nova.compute.manager [None req-25171177-b508-4c7d-af68-a3f2c99295fb tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 850.113579] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6edb1bdc-2e95-4b3f-af7c-efadf9e2d03f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.196233] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': task-1355631, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109189} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.196590] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 850.197473] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f3f8a1d-bc22-436a-a30c-8199ca3b4392 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.220412] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] b385407b-1bdd-4c53-907c-cb4c8ce16cc7/b385407b-1bdd-4c53-907c-cb4c8ce16cc7.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 850.220760] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2104b3f4-d070-495b-aff7-600fa14f024e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.243558] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Waiting for the task: (returnval){ [ 850.243558] env[61970]: value = "task-1355634" [ 850.243558] env[61970]: _type = "Task" [ 850.243558] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.250387] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': task-1355634, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.278280] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355632, 'name': Rename_Task, 'duration_secs': 0.178133} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.278510] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 850.278849] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5dae98c-cff5-48fa-baa3-1b8f841a191c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.286074] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 850.286074] env[61970]: value = "task-1355635" [ 850.286074] env[61970]: _type = "Task" [ 850.286074] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.294867] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355635, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.348867] env[61970]: INFO nova.compute.manager [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] [instance: c5c253b6-ebab-4fea-bcc4-f3bb3960070b] Took 1.03 seconds to deallocate network for instance. [ 850.383483] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355633, 'name': RelocateVM_Task} progress is 42%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.410598] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "refresh_cache-d09dbde2-5daa-4ae9-9fd5-bfbb03fba136" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.410932] env[61970]: DEBUG nova.compute.manager [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Instance network_info: |[{"id": "201e52ca-273b-4df3-b32a-d9dba51cb5c7", "address": "fa:16:3e:d1:66:5f", "network": {"id": "e02bfa6c-8f0d-4413-8dcf-e644c07792d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1410864732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9235eeeb6fa4ccf959ffbea456f3694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap201e52ca-27", "ovs_interfaceid": "201e52ca-273b-4df3-b32a-d9dba51cb5c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 850.411387] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:66:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '510d3c47-3615-43d5-aa5d-a279fd915e71', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '201e52ca-273b-4df3-b32a-d9dba51cb5c7', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 850.421216] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Creating folder: Project (c9235eeeb6fa4ccf959ffbea456f3694). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 850.422037] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-df7a8149-5afa-47a3-aa57-1b26a95c8066 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.433951] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Created folder: Project (c9235eeeb6fa4ccf959ffbea456f3694) in parent group-v288740. [ 850.434164] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Creating folder: Instances. Parent ref: group-v288786. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 850.434430] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-46fc2c63-961e-4464-9f8f-01e6e540621f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.446066] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Created folder: Instances in parent group-v288786. [ 850.446066] env[61970]: DEBUG oslo.service.loopingcall [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.446323] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 850.446473] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bb457c5a-a21e-4a5b-aa49-174ff758ccb5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.467706] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 850.467706] env[61970]: value = "task-1355638" [ 850.467706] env[61970]: _type = "Task" [ 850.467706] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.476138] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355638, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.547795] env[61970]: INFO nova.compute.manager [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Took 30.12 seconds to build instance. [ 850.579947] env[61970]: ERROR nova.scheduler.client.report [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [req-989f42cf-614b-48c0-a19e-75a644e6295d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fc0f3e4e-5ef5-45e9-8335-cc68743405aa. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-989f42cf-614b-48c0-a19e-75a644e6295d"}]} [ 850.600283] env[61970]: DEBUG nova.scheduler.client.report [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Refreshing inventories for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 850.617799] env[61970]: DEBUG nova.scheduler.client.report [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Updating ProviderTree inventory for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 850.618033] env[61970]: DEBUG nova.compute.provider_tree [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 850.622712] env[61970]: DEBUG nova.compute.manager [req-1bc4a35a-ae97-4877-9c39-0134f801aae2 req-f15c4ef1-198d-46da-b6d4-76ab14b056f5 service nova] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Received event network-changed-201e52ca-273b-4df3-b32a-d9dba51cb5c7 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.622712] env[61970]: DEBUG nova.compute.manager [req-1bc4a35a-ae97-4877-9c39-0134f801aae2 req-f15c4ef1-198d-46da-b6d4-76ab14b056f5 service nova] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Refreshing instance network info cache due to event network-changed-201e52ca-273b-4df3-b32a-d9dba51cb5c7. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 850.622712] env[61970]: DEBUG oslo_concurrency.lockutils [req-1bc4a35a-ae97-4877-9c39-0134f801aae2 req-f15c4ef1-198d-46da-b6d4-76ab14b056f5 service nova] Acquiring lock "refresh_cache-d09dbde2-5daa-4ae9-9fd5-bfbb03fba136" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.622983] env[61970]: DEBUG oslo_concurrency.lockutils [req-1bc4a35a-ae97-4877-9c39-0134f801aae2 req-f15c4ef1-198d-46da-b6d4-76ab14b056f5 service nova] Acquired lock "refresh_cache-d09dbde2-5daa-4ae9-9fd5-bfbb03fba136" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.622983] env[61970]: DEBUG nova.network.neutron [req-1bc4a35a-ae97-4877-9c39-0134f801aae2 req-f15c4ef1-198d-46da-b6d4-76ab14b056f5 service nova] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Refreshing network info cache for port 201e52ca-273b-4df3-b32a-d9dba51cb5c7 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 850.637230] env[61970]: DEBUG nova.scheduler.client.report [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Refreshing aggregate associations for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa, aggregates: None {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 850.657709] env[61970]: DEBUG nova.scheduler.client.report [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Refreshing trait associations for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 850.754019] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': task-1355634, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.797108] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355635, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.885613] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355633, 'name': RelocateVM_Task} progress is 54%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.931876] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb0eb03-f321-4f0e-bd86-36b666e4a613 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.940293] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ead6d85a-034e-49b5-adf0-8ec5d6a0cbfe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.974505] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c95ae4-e232-4df2-b0e6-d6f0a28f06a1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.988563] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3599ee7-03f2-42b7-b983-8f6ae61b1dd7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.993539] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355638, 'name': CreateVM_Task} progress is 25%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.006203] env[61970]: DEBUG nova.compute.provider_tree [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 851.049699] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06110e0d-0060-486e-8012-811db1f141ac tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "b19cb516-b163-4bed-ba5b-139a0a18fc05" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.598s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.253940] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': task-1355634, 'name': ReconfigVM_Task, 'duration_secs': 0.585204} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.256881] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Reconfigured VM instance instance-00000042 to attach disk [datastore1] b385407b-1bdd-4c53-907c-cb4c8ce16cc7/b385407b-1bdd-4c53-907c-cb4c8ce16cc7.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 851.257950] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-251b1fe0-8cc5-4b3d-bd3f-b9994f6ebabd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.266201] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Waiting for the task: (returnval){ [ 851.266201] env[61970]: value = "task-1355639" [ 851.266201] env[61970]: _type = "Task" [ 851.266201] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.280759] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': task-1355639, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.303792] env[61970]: DEBUG oslo_vmware.api [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355635, 'name': PowerOnVM_Task, 'duration_secs': 0.832165} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.303792] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 851.303792] env[61970]: INFO nova.compute.manager [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Took 10.09 seconds to spawn the instance on the hypervisor. [ 851.303792] env[61970]: DEBUG nova.compute.manager [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.305812] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2396a3a0-c4c2-4c52-bd2d-b963848a0128 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.390021] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355633, 'name': RelocateVM_Task} progress is 67%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.392119] env[61970]: INFO nova.scheduler.client.report [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Deleted allocations for instance c5c253b6-ebab-4fea-bcc4-f3bb3960070b [ 851.481691] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355638, 'name': CreateVM_Task, 'duration_secs': 0.959913} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.481930] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 851.482763] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.482932] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.483413] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.483784] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b51ca6a-7fd1-49a7-9299-a7330a3a3e14 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.489841] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 851.489841] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5297f0db-14df-74e0-5012-a29a18fa15ae" [ 851.489841] env[61970]: _type = "Task" [ 851.489841] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.497912] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5297f0db-14df-74e0-5012-a29a18fa15ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.506716] env[61970]: DEBUG nova.network.neutron [req-1bc4a35a-ae97-4877-9c39-0134f801aae2 req-f15c4ef1-198d-46da-b6d4-76ab14b056f5 service nova] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Updated VIF entry in instance network info cache for port 201e52ca-273b-4df3-b32a-d9dba51cb5c7. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 851.507089] env[61970]: DEBUG nova.network.neutron [req-1bc4a35a-ae97-4877-9c39-0134f801aae2 req-f15c4ef1-198d-46da-b6d4-76ab14b056f5 service nova] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Updating instance_info_cache with network_info: [{"id": "201e52ca-273b-4df3-b32a-d9dba51cb5c7", "address": "fa:16:3e:d1:66:5f", "network": {"id": "e02bfa6c-8f0d-4413-8dcf-e644c07792d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1410864732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9235eeeb6fa4ccf959ffbea456f3694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap201e52ca-27", "ovs_interfaceid": "201e52ca-273b-4df3-b32a-d9dba51cb5c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.544093] env[61970]: DEBUG nova.scheduler.client.report [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Updated inventory for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with generation 84 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 851.544532] env[61970]: DEBUG nova.compute.provider_tree [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Updating resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa generation from 84 to 85 during operation: update_inventory {{(pid=61970) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 851.544682] env[61970]: DEBUG nova.compute.provider_tree [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 851.552630] env[61970]: DEBUG nova.compute.manager [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 851.777149] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': task-1355639, 'name': Rename_Task, 'duration_secs': 0.216935} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.777808] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 851.778294] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5a1b826-2524-41bd-a1ed-34fee8db2e02 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.787367] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Waiting for the task: (returnval){ [ 851.787367] env[61970]: value = "task-1355640" [ 851.787367] env[61970]: _type = "Task" [ 851.787367] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.799700] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': task-1355640, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.828516] env[61970]: INFO nova.compute.manager [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Took 29.81 seconds to build instance. [ 851.887241] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355633, 'name': RelocateVM_Task} progress is 82%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.899993] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7548be9-d3cc-4bef-bdaf-90b9e2984b35 tempest-ServerRescueTestJSONUnderV235-840084764 tempest-ServerRescueTestJSONUnderV235-840084764-project-member] Lock "c5c253b6-ebab-4fea-bcc4-f3bb3960070b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.786s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.001504] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5297f0db-14df-74e0-5012-a29a18fa15ae, 'name': SearchDatastore_Task, 'duration_secs': 0.090433} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.001868] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.002187] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.002380] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.002574] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.002795] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.003120] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-74745de4-ec3b-4dac-9689-0fb3973ef793 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.009996] env[61970]: DEBUG oslo_concurrency.lockutils [req-1bc4a35a-ae97-4877-9c39-0134f801aae2 req-f15c4ef1-198d-46da-b6d4-76ab14b056f5 service nova] Releasing lock "refresh_cache-d09dbde2-5daa-4ae9-9fd5-bfbb03fba136" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.013592] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 852.013778] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 852.014648] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03ed45ea-b467-45b2-9125-5cd4624f1fca {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.020502] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 852.020502] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528bc77c-7ce8-861c-832e-80891be3ca01" [ 852.020502] env[61970]: _type = "Task" [ 852.020502] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.028738] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528bc77c-7ce8-861c-832e-80891be3ca01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.050278] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.859s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.052680] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.806s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.054315] env[61970]: INFO nova.compute.claims [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 852.072837] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.073881] env[61970]: INFO nova.scheduler.client.report [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Deleted allocations for instance a78b63a5-6bb8-4271-90d8-1e86fb29db4f [ 852.297969] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': task-1355640, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.329869] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a35f3753-71a9-4406-84b5-40fb5e974062 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.359s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.388314] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355633, 'name': RelocateVM_Task} progress is 97%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.404296] env[61970]: DEBUG nova.compute.manager [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 852.490465] env[61970]: DEBUG oslo_concurrency.lockutils [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "b19cb516-b163-4bed-ba5b-139a0a18fc05" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.490465] env[61970]: DEBUG oslo_concurrency.lockutils [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "b19cb516-b163-4bed-ba5b-139a0a18fc05" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.531599] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528bc77c-7ce8-861c-832e-80891be3ca01, 'name': SearchDatastore_Task, 'duration_secs': 0.091602} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.532617] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31358718-e569-4e69-acf6-8ebf3fd0f4bb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.537795] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 852.537795] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526db575-e6e9-86c6-10db-920b6fe7f6ff" [ 852.537795] env[61970]: _type = "Task" [ 852.537795] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.545439] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526db575-e6e9-86c6-10db-920b6fe7f6ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.582315] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62cf8311-4047-4ac9-a028-41f1c2393c63 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "a78b63a5-6bb8-4271-90d8-1e86fb29db4f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.332s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.799267] env[61970]: DEBUG oslo_vmware.api [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': task-1355640, 'name': PowerOnVM_Task, 'duration_secs': 0.659877} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.800423] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 852.800423] env[61970]: INFO nova.compute.manager [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Took 9.18 seconds to spawn the instance on the hypervisor. [ 852.800423] env[61970]: DEBUG nova.compute.manager [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.800665] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f2c2490-b789-4f54-a24c-4153b48349fc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.832656] env[61970]: DEBUG nova.compute.manager [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 852.889419] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355633, 'name': RelocateVM_Task} progress is 97%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.927808] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.993745] env[61970]: DEBUG nova.compute.utils [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 853.048080] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526db575-e6e9-86c6-10db-920b6fe7f6ff, 'name': SearchDatastore_Task, 'duration_secs': 0.02449} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.048384] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.048533] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] d09dbde2-5daa-4ae9-9fd5-bfbb03fba136/d09dbde2-5daa-4ae9-9fd5-bfbb03fba136.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 853.048779] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b1200878-4c61-49f6-80af-252aa0a0ffbe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.055413] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 853.055413] env[61970]: value = "task-1355641" [ 853.055413] env[61970]: _type = "Task" [ 853.055413] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.062869] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355641, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.278591] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc3175cb-5d64-4054-afc9-fd3c02890b5e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.286937] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32403a0a-db4d-4517-8ccc-ed19c8695937 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.324716] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3b224f-0968-4fa3-912f-94b77559b083 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.326814] env[61970]: INFO nova.compute.manager [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Took 29.34 seconds to build instance. [ 853.331272] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559da658-a870-41ee-acd0-db2d3f19d96d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.347652] env[61970]: DEBUG nova.compute.provider_tree [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.362757] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.388712] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355633, 'name': RelocateVM_Task} progress is 97%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.497107] env[61970]: DEBUG oslo_concurrency.lockutils [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "b19cb516-b163-4bed-ba5b-139a0a18fc05" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.566112] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355641, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.830690] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8d376016-f0c3-4087-ad3c-ba94eaef9288 tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Lock "b385407b-1bdd-4c53-907c-cb4c8ce16cc7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.664s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.853620] env[61970]: DEBUG nova.scheduler.client.report [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 853.891047] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355633, 'name': RelocateVM_Task} progress is 98%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.007210] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 854.009293] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 854.076023] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355641, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.983502} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.076023] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] d09dbde2-5daa-4ae9-9fd5-bfbb03fba136/d09dbde2-5daa-4ae9-9fd5-bfbb03fba136.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 854.076023] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.076023] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a0ee0d9d-cc7a-46d9-ba1f-65faecd8429a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.092153] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 854.092153] env[61970]: value = "task-1355642" [ 854.092153] env[61970]: _type = "Task" [ 854.092153] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.104841] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355642, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.112111] env[61970]: DEBUG nova.compute.manager [req-292ced1f-ffb0-4363-8488-28f7db7f8c52 req-f9e648b5-0a77-44ec-a0c6-223b096c9bcb service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Received event network-changed-64a3e684-f5d3-4dd2-b8c9-dc46565f71bd {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.112111] env[61970]: DEBUG nova.compute.manager [req-292ced1f-ffb0-4363-8488-28f7db7f8c52 req-f9e648b5-0a77-44ec-a0c6-223b096c9bcb service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Refreshing instance network info cache due to event network-changed-64a3e684-f5d3-4dd2-b8c9-dc46565f71bd. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 854.112111] env[61970]: DEBUG oslo_concurrency.lockutils [req-292ced1f-ffb0-4363-8488-28f7db7f8c52 req-f9e648b5-0a77-44ec-a0c6-223b096c9bcb service nova] Acquiring lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.112111] env[61970]: DEBUG oslo_concurrency.lockutils [req-292ced1f-ffb0-4363-8488-28f7db7f8c52 req-f9e648b5-0a77-44ec-a0c6-223b096c9bcb service nova] Acquired lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.112806] env[61970]: DEBUG nova.network.neutron [req-292ced1f-ffb0-4363-8488-28f7db7f8c52 req-f9e648b5-0a77-44ec-a0c6-223b096c9bcb service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Refreshing network info cache for port 64a3e684-f5d3-4dd2-b8c9-dc46565f71bd {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 854.334268] env[61970]: DEBUG nova.compute.manager [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 854.359091] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.306s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.359387] env[61970]: DEBUG nova.compute.manager [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 854.363238] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.519s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.363238] env[61970]: INFO nova.compute.claims [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 854.366941] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.367175] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.367613] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.367811] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.368250] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.370645] env[61970]: INFO nova.compute.manager [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Terminating instance [ 854.372373] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "refresh_cache-7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.372521] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquired lock "refresh_cache-7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.372681] env[61970]: DEBUG nova.network.neutron [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 854.392897] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355633, 'name': RelocateVM_Task, 'duration_secs': 4.267014} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.392897] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Volume attach. Driver type: vmdk {{(pid=61970) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 854.392897] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288764', 'volume_id': '4020284d-a273-459a-bb8c-84753ab98b1e', 'name': 'volume-4020284d-a273-459a-bb8c-84753ab98b1e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6838c6a4-8241-4b6f-ad58-3acfd7afdbf4', 'attached_at': '', 'detached_at': '', 'volume_id': '4020284d-a273-459a-bb8c-84753ab98b1e', 'serial': '4020284d-a273-459a-bb8c-84753ab98b1e'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 854.393251] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc146521-c0f0-4517-b50f-cbb86a41c08a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.415376] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ed3431-2ba9-4481-a883-5c5218bfc6f6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.442463] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] volume-4020284d-a273-459a-bb8c-84753ab98b1e/volume-4020284d-a273-459a-bb8c-84753ab98b1e.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.443047] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3607fcc-1012-4153-9b3b-52ca5d48dbe8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.466040] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Waiting for the task: (returnval){ [ 854.466040] env[61970]: value = "task-1355643" [ 854.466040] env[61970]: _type = "Task" [ 854.466040] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.475799] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355643, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.515632] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 854.515823] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Starting heal instance info cache {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 854.516980] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Rebuilding the list of instances to heal {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 854.593305] env[61970]: DEBUG oslo_concurrency.lockutils [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "b19cb516-b163-4bed-ba5b-139a0a18fc05" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.593571] env[61970]: DEBUG oslo_concurrency.lockutils [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "b19cb516-b163-4bed-ba5b-139a0a18fc05" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.593810] env[61970]: INFO nova.compute.manager [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Attaching volume f1e6410e-b97a-4846-b2db-fcd81bc4e2de to /dev/sdb [ 854.605719] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355642, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064107} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.605999] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.606824] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c96a3a-e851-4833-8c63-35241ea80c99 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.633547] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] d09dbde2-5daa-4ae9-9fd5-bfbb03fba136/d09dbde2-5daa-4ae9-9fd5-bfbb03fba136.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.634652] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f87c9142-597b-4498-a05d-33f4753b9ec1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.654107] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03e96654-20d7-4c02-af0a-b35b07a48d8c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.663100] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb080eed-087f-4284-bde6-1f16138073fe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.665803] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 854.665803] env[61970]: value = "task-1355644" [ 854.665803] env[61970]: _type = "Task" [ 854.665803] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.676291] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355644, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.680282] env[61970]: DEBUG nova.virt.block_device [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Updating existing volume attachment record: 18c6eca9-acba-48cb-8bec-5a32b08a719a {{(pid=61970) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 854.751137] env[61970]: DEBUG nova.compute.manager [req-7b25af86-c0e9-4740-9973-ce115e5b8dc6 req-b74ef977-97ed-4748-bc81-3f16483b808f service nova] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Received event network-changed-6875c46d-b7c5-4912-a1f0-932b3ab900d8 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.751333] env[61970]: DEBUG nova.compute.manager [req-7b25af86-c0e9-4740-9973-ce115e5b8dc6 req-b74ef977-97ed-4748-bc81-3f16483b808f service nova] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Refreshing instance network info cache due to event network-changed-6875c46d-b7c5-4912-a1f0-932b3ab900d8. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 854.751534] env[61970]: DEBUG oslo_concurrency.lockutils [req-7b25af86-c0e9-4740-9973-ce115e5b8dc6 req-b74ef977-97ed-4748-bc81-3f16483b808f service nova] Acquiring lock "refresh_cache-b385407b-1bdd-4c53-907c-cb4c8ce16cc7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.751666] env[61970]: DEBUG oslo_concurrency.lockutils [req-7b25af86-c0e9-4740-9973-ce115e5b8dc6 req-b74ef977-97ed-4748-bc81-3f16483b808f service nova] Acquired lock "refresh_cache-b385407b-1bdd-4c53-907c-cb4c8ce16cc7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.751964] env[61970]: DEBUG nova.network.neutron [req-7b25af86-c0e9-4740-9973-ce115e5b8dc6 req-b74ef977-97ed-4748-bc81-3f16483b808f service nova] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Refreshing network info cache for port 6875c46d-b7c5-4912-a1f0-932b3ab900d8 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 854.867629] env[61970]: DEBUG oslo_concurrency.lockutils [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.869087] env[61970]: DEBUG nova.compute.utils [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 854.870593] env[61970]: DEBUG nova.compute.manager [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Not allocating networking since 'none' was specified. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 854.912754] env[61970]: DEBUG nova.network.neutron [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 854.939563] env[61970]: DEBUG nova.network.neutron [req-292ced1f-ffb0-4363-8488-28f7db7f8c52 req-f9e648b5-0a77-44ec-a0c6-223b096c9bcb service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Updated VIF entry in instance network info cache for port 64a3e684-f5d3-4dd2-b8c9-dc46565f71bd. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 854.940113] env[61970]: DEBUG nova.network.neutron [req-292ced1f-ffb0-4363-8488-28f7db7f8c52 req-f9e648b5-0a77-44ec-a0c6-223b096c9bcb service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Updating instance_info_cache with network_info: [{"id": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "address": "fa:16:3e:14:8c:ec", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64a3e684-f5", "ovs_interfaceid": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.981581] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355643, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.001132] env[61970]: DEBUG nova.network.neutron [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.022457] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Skipping network cache update for instance because it is being deleted. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 855.022755] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 855.022948] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 855.023157] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 855.023333] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 855.090087] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "refresh_cache-5d4d97e1-5812-4884-a1ea-6eb0cee591c0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.090481] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquired lock "refresh_cache-5d4d97e1-5812-4884-a1ea-6eb0cee591c0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.090481] env[61970]: DEBUG nova.network.neutron [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Forcefully refreshing network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 855.090697] env[61970]: DEBUG nova.objects.instance [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lazy-loading 'info_cache' on Instance uuid 5d4d97e1-5812-4884-a1ea-6eb0cee591c0 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 855.181249] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355644, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.373806] env[61970]: DEBUG nova.compute.manager [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 855.443213] env[61970]: DEBUG oslo_concurrency.lockutils [req-292ced1f-ffb0-4363-8488-28f7db7f8c52 req-f9e648b5-0a77-44ec-a0c6-223b096c9bcb service nova] Releasing lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.477475] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "e2f51299-56fe-46cc-9e08-2246178b9db7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.477761] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "e2f51299-56fe-46cc-9e08-2246178b9db7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.477979] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "e2f51299-56fe-46cc-9e08-2246178b9db7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.478172] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "e2f51299-56fe-46cc-9e08-2246178b9db7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.478335] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "e2f51299-56fe-46cc-9e08-2246178b9db7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.479921] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355643, 'name': ReconfigVM_Task, 'duration_secs': 0.675758} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.480648] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Reconfigured VM instance instance-00000043 to attach disk [datastore1] volume-4020284d-a273-459a-bb8c-84753ab98b1e/volume-4020284d-a273-459a-bb8c-84753ab98b1e.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.487613] env[61970]: INFO nova.compute.manager [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Terminating instance [ 855.491054] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dea55928-0a26-4179-8bd2-cabeb62eee22 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.502301] env[61970]: DEBUG nova.compute.manager [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 855.502301] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 855.504603] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d711129f-5614-4c44-ad35-2ecb16590962 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.507285] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Releasing lock "refresh_cache-7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.507677] env[61970]: DEBUG nova.compute.manager [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 855.507864] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 855.508614] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67c68fe-b302-4a82-aa91-7fa5d753fb08 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.514504] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Waiting for the task: (returnval){ [ 855.514504] env[61970]: value = "task-1355648" [ 855.514504] env[61970]: _type = "Task" [ 855.514504] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.518888] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 855.526215] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7f18a427-22b1-46b9-bbd7-658fed892aa1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.534729] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355648, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.537858] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 855.538227] env[61970]: DEBUG oslo_vmware.api [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 855.538227] env[61970]: value = "task-1355649" [ 855.538227] env[61970]: _type = "Task" [ 855.538227] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.541618] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee8c2ccc-a6d8-4446-9723-906b945b9bfb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.553299] env[61970]: DEBUG oslo_vmware.api [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355649, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.557769] env[61970]: DEBUG oslo_vmware.api [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 855.557769] env[61970]: value = "task-1355650" [ 855.557769] env[61970]: _type = "Task" [ 855.557769] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.566546] env[61970]: DEBUG oslo_vmware.api [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355650, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.681099] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355644, 'name': ReconfigVM_Task, 'duration_secs': 0.576616} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.682095] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Reconfigured VM instance instance-00000044 to attach disk [datastore2] d09dbde2-5daa-4ae9-9fd5-bfbb03fba136/d09dbde2-5daa-4ae9-9fd5-bfbb03fba136.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.682095] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f32e4151-ddef-4fff-ad60-23973589aa31 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.688182] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fc3f26d-3aed-48ac-becc-cb3cd6d3321f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.693404] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 855.693404] env[61970]: value = "task-1355651" [ 855.693404] env[61970]: _type = "Task" [ 855.693404] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.704218] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc011142-b4d3-43ed-bfc3-9e348b3faa00 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.712533] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355651, 'name': Rename_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.747448] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17b913bb-370a-4117-9198-4c4613dff15d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.756658] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ddb1978-7a54-42b4-a3c4-8d6aea6930a6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.763991] env[61970]: DEBUG nova.network.neutron [req-7b25af86-c0e9-4740-9973-ce115e5b8dc6 req-b74ef977-97ed-4748-bc81-3f16483b808f service nova] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Updated VIF entry in instance network info cache for port 6875c46d-b7c5-4912-a1f0-932b3ab900d8. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 855.764406] env[61970]: DEBUG nova.network.neutron [req-7b25af86-c0e9-4740-9973-ce115e5b8dc6 req-b74ef977-97ed-4748-bc81-3f16483b808f service nova] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Updating instance_info_cache with network_info: [{"id": "6875c46d-b7c5-4912-a1f0-932b3ab900d8", "address": "fa:16:3e:00:b6:23", "network": {"id": "01909a13-6915-4607-b08d-9d5aac1ab78f", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1225281584-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.237", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2c5df141b77042c28368dc4474f7d917", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "163e60bd-32d6-41c5-95e6-2eb10c5c9245", "external-id": "nsx-vlan-transportzone-716", "segmentation_id": 716, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6875c46d-b7", "ovs_interfaceid": "6875c46d-b7c5-4912-a1f0-932b3ab900d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.776060] env[61970]: DEBUG nova.compute.provider_tree [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.030063] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355648, 'name': ReconfigVM_Task, 'duration_secs': 0.22541} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.030384] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288764', 'volume_id': '4020284d-a273-459a-bb8c-84753ab98b1e', 'name': 'volume-4020284d-a273-459a-bb8c-84753ab98b1e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6838c6a4-8241-4b6f-ad58-3acfd7afdbf4', 'attached_at': '', 'detached_at': '', 'volume_id': '4020284d-a273-459a-bb8c-84753ab98b1e', 'serial': '4020284d-a273-459a-bb8c-84753ab98b1e'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 856.030913] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7070438a-2e03-4fbe-98af-010da325d5b8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.037245] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Waiting for the task: (returnval){ [ 856.037245] env[61970]: value = "task-1355652" [ 856.037245] env[61970]: _type = "Task" [ 856.037245] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.045494] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355652, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.052731] env[61970]: DEBUG oslo_vmware.api [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355649, 'name': PowerOffVM_Task, 'duration_secs': 0.295479} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.052969] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 856.053155] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 856.053382] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd9f7dfd-0e54-40db-aeb0-0f5a297e07e1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.066474] env[61970]: DEBUG oslo_vmware.api [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355650, 'name': PowerOffVM_Task, 'duration_secs': 0.307552} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.066633] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 856.066984] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 856.066984] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b01dbf41-0e0c-45bf-a204-be6c9a2d97bb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.093137] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 856.093533] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 856.093533] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Deleting the datastore file [datastore2] 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 856.093870] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8e580f5b-b9dc-432b-b3cc-e5d412b8d29f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.101964] env[61970]: DEBUG oslo_vmware.api [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for the task: (returnval){ [ 856.101964] env[61970]: value = "task-1355655" [ 856.101964] env[61970]: _type = "Task" [ 856.101964] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.110814] env[61970]: DEBUG oslo_vmware.api [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355655, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.117019] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 856.117236] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 856.117424] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Deleting the datastore file [datastore2] e2f51299-56fe-46cc-9e08-2246178b9db7 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 856.117691] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14f9bc4e-6f33-4198-a345-d37ea6f454ce {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.123551] env[61970]: DEBUG oslo_vmware.api [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 856.123551] env[61970]: value = "task-1355656" [ 856.123551] env[61970]: _type = "Task" [ 856.123551] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.131302] env[61970]: DEBUG oslo_vmware.api [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355656, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.205803] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355651, 'name': Rename_Task, 'duration_secs': 0.307453} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.206554] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.207029] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-23f8d3ca-a3cd-4f23-aba9-bd92327c154f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.213761] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 856.213761] env[61970]: value = "task-1355657" [ 856.213761] env[61970]: _type = "Task" [ 856.213761] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.223061] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355657, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.268636] env[61970]: DEBUG oslo_concurrency.lockutils [req-7b25af86-c0e9-4740-9973-ce115e5b8dc6 req-b74ef977-97ed-4748-bc81-3f16483b808f service nova] Releasing lock "refresh_cache-b385407b-1bdd-4c53-907c-cb4c8ce16cc7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.280039] env[61970]: DEBUG nova.scheduler.client.report [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.390219] env[61970]: DEBUG nova.compute.manager [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 856.418662] env[61970]: DEBUG nova.virt.hardware [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 856.419076] env[61970]: DEBUG nova.virt.hardware [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 856.419335] env[61970]: DEBUG nova.virt.hardware [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.419620] env[61970]: DEBUG nova.virt.hardware [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 856.419855] env[61970]: DEBUG nova.virt.hardware [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.420138] env[61970]: DEBUG nova.virt.hardware [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 856.420481] env[61970]: DEBUG nova.virt.hardware [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 856.420745] env[61970]: DEBUG nova.virt.hardware [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 856.421035] env[61970]: DEBUG nova.virt.hardware [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 856.421326] env[61970]: DEBUG nova.virt.hardware [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 856.421641] env[61970]: DEBUG nova.virt.hardware [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 856.422964] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9496b167-fa19-48fd-b9b8-d02e51cd5341 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.434508] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904f2aec-aeaf-46ae-9b61-a57f861bf4bc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.455627] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Instance VIF info [] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 856.465542] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Creating folder: Project (93f97be6a44a4d0fa251dd5614c92fbe). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 856.466016] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-726ec547-4ba2-4b84-ad12-60ce79337e42 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.476210] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Created folder: Project (93f97be6a44a4d0fa251dd5614c92fbe) in parent group-v288740. [ 856.476544] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Creating folder: Instances. Parent ref: group-v288791. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 856.476900] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b485ecec-b8ff-4f3e-b123-d03f88c9b82c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.488662] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Created folder: Instances in parent group-v288791. [ 856.489307] env[61970]: DEBUG oslo.service.loopingcall [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.489601] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 856.489930] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e28745c6-8188-46a7-aa12-9f28a22ce7f4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.516622] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 856.516622] env[61970]: value = "task-1355660" [ 856.516622] env[61970]: _type = "Task" [ 856.516622] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.527784] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355660, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.549756] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355652, 'name': Rename_Task, 'duration_secs': 0.188918} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.550190] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.550537] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2885e52-3616-42f3-901c-d100320cd129 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.557549] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Waiting for the task: (returnval){ [ 856.557549] env[61970]: value = "task-1355661" [ 856.557549] env[61970]: _type = "Task" [ 856.557549] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.570216] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355661, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.615818] env[61970]: DEBUG oslo_vmware.api [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Task: {'id': task-1355655, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131633} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.616074] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 856.616260] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 856.616451] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 856.616676] env[61970]: INFO nova.compute.manager [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Took 1.11 seconds to destroy the instance on the hypervisor. [ 856.616937] env[61970]: DEBUG oslo.service.loopingcall [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.617148] env[61970]: DEBUG nova.compute.manager [-] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 856.617242] env[61970]: DEBUG nova.network.neutron [-] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 856.632598] env[61970]: DEBUG oslo_vmware.api [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355656, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.300258} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.632852] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 856.633038] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 856.633232] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 856.633401] env[61970]: INFO nova.compute.manager [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Took 1.13 seconds to destroy the instance on the hypervisor. [ 856.633768] env[61970]: DEBUG oslo.service.loopingcall [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.633849] env[61970]: DEBUG nova.compute.manager [-] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 856.633945] env[61970]: DEBUG nova.network.neutron [-] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 856.636099] env[61970]: DEBUG nova.network.neutron [-] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 856.723389] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355657, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.784436] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.422s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.788123] env[61970]: DEBUG nova.compute.manager [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 856.788123] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.132s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.789212] env[61970]: INFO nova.compute.claims [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 856.883555] env[61970]: DEBUG nova.network.neutron [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Updating instance_info_cache with network_info: [{"id": "1080e2f3-c9a3-48bc-bd65-b5f231bacdc7", "address": "fa:16:3e:55:7f:ba", "network": {"id": "700300d5-dfe3-4ff7-86ed-43a4f2c15818", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1754634816-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e91895bd10c74f8d9a101274d87cdc71", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ba56987-9dc3-4c76-a4e2-942b05355bdb", "external-id": "nsx-vlan-transportzone-698", "segmentation_id": 698, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1080e2f3-c9", "ovs_interfaceid": "1080e2f3-c9a3-48bc-bd65-b5f231bacdc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.026828] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355660, 'name': CreateVM_Task} progress is 25%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.067602] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355661, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.118025] env[61970]: DEBUG nova.compute.manager [req-a56ee484-0927-401b-b79c-a9a528acd6ca req-ef241c5c-f630-47c7-a3c9-8c6774d46986 service nova] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Received event network-vif-deleted-817c117f-93e2-414c-ac36-01fbf658b9d1 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.118025] env[61970]: INFO nova.compute.manager [req-a56ee484-0927-401b-b79c-a9a528acd6ca req-ef241c5c-f630-47c7-a3c9-8c6774d46986 service nova] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Neutron deleted interface 817c117f-93e2-414c-ac36-01fbf658b9d1; detaching it from the instance and deleting it from the info cache [ 857.118025] env[61970]: DEBUG nova.network.neutron [req-a56ee484-0927-401b-b79c-a9a528acd6ca req-ef241c5c-f630-47c7-a3c9-8c6774d46986 service nova] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.138973] env[61970]: DEBUG nova.network.neutron [-] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.229040] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355657, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.298078] env[61970]: DEBUG nova.compute.utils [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 857.299798] env[61970]: DEBUG nova.compute.manager [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 857.299978] env[61970]: DEBUG nova.network.neutron [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 857.357362] env[61970]: DEBUG nova.policy [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7bb7fac10cbf4876868fd5858e18d34a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1458b73cd8a4356816756e4c7d05925', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 857.386496] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Releasing lock "refresh_cache-5d4d97e1-5812-4884-a1ea-6eb0cee591c0" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.386676] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Updated the network info_cache for instance {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 857.386864] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 857.387033] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 857.387201] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 857.387343] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 857.387599] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 857.387781] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 857.387923] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61970) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 857.388080] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 857.530051] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355660, 'name': CreateVM_Task, 'duration_secs': 0.94234} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.530270] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 857.531045] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.531229] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.531635] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 857.531902] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80a99cf1-30ba-4c75-8a99-6fe425475385 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.538958] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 857.538958] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52bfefd3-1dfd-86f5-7ab8-4a591da08289" [ 857.538958] env[61970]: _type = "Task" [ 857.538958] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.547684] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52bfefd3-1dfd-86f5-7ab8-4a591da08289, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.567859] env[61970]: DEBUG oslo_vmware.api [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355661, 'name': PowerOnVM_Task, 'duration_secs': 0.757036} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.568152] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.568372] env[61970]: INFO nova.compute.manager [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Took 9.82 seconds to spawn the instance on the hypervisor. [ 857.568552] env[61970]: DEBUG nova.compute.manager [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.569355] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb14ebf4-0d93-46c6-bc1d-fc5f174876ad {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.597447] env[61970]: DEBUG nova.network.neutron [-] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.621672] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0bf3a322-a7bc-4caf-98fe-0ecb5f54545d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.631937] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05a65ea4-779b-4feb-aa05-874463530870 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.643269] env[61970]: INFO nova.compute.manager [-] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Took 1.03 seconds to deallocate network for instance. [ 857.665047] env[61970]: DEBUG nova.compute.manager [req-a56ee484-0927-401b-b79c-a9a528acd6ca req-ef241c5c-f630-47c7-a3c9-8c6774d46986 service nova] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Detach interface failed, port_id=817c117f-93e2-414c-ac36-01fbf658b9d1, reason: Instance e2f51299-56fe-46cc-9e08-2246178b9db7 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 857.684581] env[61970]: DEBUG nova.network.neutron [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Successfully created port: 9096db72-8476-4d1f-8fcc-cc5ddd54d61d {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 857.725202] env[61970]: DEBUG oslo_vmware.api [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355657, 'name': PowerOnVM_Task, 'duration_secs': 1.417156} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.725682] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.725682] env[61970]: INFO nova.compute.manager [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Took 9.37 seconds to spawn the instance on the hypervisor. [ 857.725910] env[61970]: DEBUG nova.compute.manager [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.726811] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ee1613-7512-4c79-a4a7-f0ee333f0235 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.800718] env[61970]: DEBUG nova.compute.manager [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 857.891425] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.049024] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52bfefd3-1dfd-86f5-7ab8-4a591da08289, 'name': SearchDatastore_Task, 'duration_secs': 0.012052} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.051712] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.052012] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 858.052280] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.053328] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.053328] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 858.053584] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1130c52b-19e6-4773-a8fd-793b57d731b4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.068889] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 858.069322] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 858.070594] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac7448c6-ddf3-448f-bec2-7d0481bf5e4c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.084261] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 858.084261] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52c8ba61-76bc-60cf-8f5c-b594789543f4" [ 858.084261] env[61970]: _type = "Task" [ 858.084261] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.091900] env[61970]: INFO nova.compute.manager [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Took 32.22 seconds to build instance. [ 858.104484] env[61970]: INFO nova.compute.manager [-] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Took 1.47 seconds to deallocate network for instance. [ 858.104858] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52c8ba61-76bc-60cf-8f5c-b594789543f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.139469] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-084c5dfd-d3a1-4d17-a37c-8583d3b6a56c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.148836] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f426df-8ea4-48bb-bbc3-c14743487195 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.153558] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.190720] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8571df04-e7e9-4427-9213-df98adbc13bf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.199556] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f9f584d-19f3-4fa9-b56b-d745874b4600 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.215227] env[61970]: DEBUG nova.compute.provider_tree [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.244474] env[61970]: INFO nova.compute.manager [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Took 30.44 seconds to build instance. [ 858.595758] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f3413a27-f1cd-4468-a748-7def00ad049a tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Lock "6838c6a4-8241-4b6f-ad58-3acfd7afdbf4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.213s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.596047] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52c8ba61-76bc-60cf-8f5c-b594789543f4, 'name': SearchDatastore_Task, 'duration_secs': 0.028387} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.596815] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5915987-1b0e-4e2c-abcf-9941d806a5bd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.602442] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 858.602442] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]520cc850-eb07-4660-f1ce-03273a821492" [ 858.602442] env[61970]: _type = "Task" [ 858.602442] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.610298] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]520cc850-eb07-4660-f1ce-03273a821492, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.615385] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.722019] env[61970]: DEBUG nova.scheduler.client.report [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 858.746972] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fd9671a6-ae52-40e5-b62d-275beef8ed0a tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "d09dbde2-5daa-4ae9-9fd5-bfbb03fba136" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.791s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.791866] env[61970]: DEBUG nova.compute.manager [req-6ee272ff-027d-4b0a-b581-9f317723e515 req-16d65c6c-81cf-4ee8-b4ed-cc3b6d42893b service nova] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Received event network-changed-1dfb626c-a7e0-455e-bffd-3eabada94ae1 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.792082] env[61970]: DEBUG nova.compute.manager [req-6ee272ff-027d-4b0a-b581-9f317723e515 req-16d65c6c-81cf-4ee8-b4ed-cc3b6d42893b service nova] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Refreshing instance network info cache due to event network-changed-1dfb626c-a7e0-455e-bffd-3eabada94ae1. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 858.792304] env[61970]: DEBUG oslo_concurrency.lockutils [req-6ee272ff-027d-4b0a-b581-9f317723e515 req-16d65c6c-81cf-4ee8-b4ed-cc3b6d42893b service nova] Acquiring lock "refresh_cache-6838c6a4-8241-4b6f-ad58-3acfd7afdbf4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.792446] env[61970]: DEBUG oslo_concurrency.lockutils [req-6ee272ff-027d-4b0a-b581-9f317723e515 req-16d65c6c-81cf-4ee8-b4ed-cc3b6d42893b service nova] Acquired lock "refresh_cache-6838c6a4-8241-4b6f-ad58-3acfd7afdbf4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.792606] env[61970]: DEBUG nova.network.neutron [req-6ee272ff-027d-4b0a-b581-9f317723e515 req-16d65c6c-81cf-4ee8-b4ed-cc3b6d42893b service nova] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Refreshing network info cache for port 1dfb626c-a7e0-455e-bffd-3eabada94ae1 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 858.810656] env[61970]: DEBUG nova.compute.manager [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 858.841209] env[61970]: DEBUG nova.virt.hardware [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 858.842134] env[61970]: DEBUG nova.virt.hardware [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 858.842134] env[61970]: DEBUG nova.virt.hardware [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 858.842347] env[61970]: DEBUG nova.virt.hardware [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 858.842648] env[61970]: DEBUG nova.virt.hardware [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 858.842916] env[61970]: DEBUG nova.virt.hardware [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 858.843296] env[61970]: DEBUG nova.virt.hardware [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 858.843582] env[61970]: DEBUG nova.virt.hardware [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 858.843881] env[61970]: DEBUG nova.virt.hardware [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 858.844184] env[61970]: DEBUG nova.virt.hardware [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 858.844544] env[61970]: DEBUG nova.virt.hardware [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.846274] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd28d105-67f1-4a73-867f-3a33ea52b888 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.857216] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa5c6d02-dcfb-4ba4-be55-b1f60c95ad44 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.026965] env[61970]: DEBUG oslo_concurrency.lockutils [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "d09dbde2-5daa-4ae9-9fd5-bfbb03fba136" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.027219] env[61970]: DEBUG oslo_concurrency.lockutils [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "d09dbde2-5daa-4ae9-9fd5-bfbb03fba136" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.027425] env[61970]: DEBUG oslo_concurrency.lockutils [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "d09dbde2-5daa-4ae9-9fd5-bfbb03fba136-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.027611] env[61970]: DEBUG oslo_concurrency.lockutils [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "d09dbde2-5daa-4ae9-9fd5-bfbb03fba136-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.027784] env[61970]: DEBUG oslo_concurrency.lockutils [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "d09dbde2-5daa-4ae9-9fd5-bfbb03fba136-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.030277] env[61970]: INFO nova.compute.manager [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Terminating instance [ 859.032158] env[61970]: DEBUG nova.compute.manager [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 859.032367] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 859.033282] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd8a4645-849f-4231-849a-8f1fd2182e4f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.040934] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 859.041072] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1146c85-3c3a-45bd-aace-342117f13022 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.047699] env[61970]: DEBUG oslo_vmware.api [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 859.047699] env[61970]: value = "task-1355663" [ 859.047699] env[61970]: _type = "Task" [ 859.047699] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.056482] env[61970]: DEBUG oslo_vmware.api [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355663, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.117350] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]520cc850-eb07-4660-f1ce-03273a821492, 'name': SearchDatastore_Task, 'duration_secs': 0.024634} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.117797] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.118231] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] e1415314-f137-4ee3-a065-3d875ef9e7ff/e1415314-f137-4ee3-a065-3d875ef9e7ff.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 859.118600] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1d70925-5875-4139-9b7a-6d02d2d499ab {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.126548] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 859.126548] env[61970]: value = "task-1355664" [ 859.126548] env[61970]: _type = "Task" [ 859.126548] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.136138] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355664, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.211923] env[61970]: DEBUG nova.compute.manager [req-d904458d-45e0-46eb-813e-bf51f1ee8cc1 req-7f887e8f-6025-4d72-b250-c4c567f3b21a service nova] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Received event network-vif-plugged-9096db72-8476-4d1f-8fcc-cc5ddd54d61d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.212249] env[61970]: DEBUG oslo_concurrency.lockutils [req-d904458d-45e0-46eb-813e-bf51f1ee8cc1 req-7f887e8f-6025-4d72-b250-c4c567f3b21a service nova] Acquiring lock "035fb797-21c1-4af9-8e66-deee8e9d083a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.212389] env[61970]: DEBUG oslo_concurrency.lockutils [req-d904458d-45e0-46eb-813e-bf51f1ee8cc1 req-7f887e8f-6025-4d72-b250-c4c567f3b21a service nova] Lock "035fb797-21c1-4af9-8e66-deee8e9d083a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.212549] env[61970]: DEBUG oslo_concurrency.lockutils [req-d904458d-45e0-46eb-813e-bf51f1ee8cc1 req-7f887e8f-6025-4d72-b250-c4c567f3b21a service nova] Lock "035fb797-21c1-4af9-8e66-deee8e9d083a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.212720] env[61970]: DEBUG nova.compute.manager [req-d904458d-45e0-46eb-813e-bf51f1ee8cc1 req-7f887e8f-6025-4d72-b250-c4c567f3b21a service nova] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] No waiting events found dispatching network-vif-plugged-9096db72-8476-4d1f-8fcc-cc5ddd54d61d {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 859.212954] env[61970]: WARNING nova.compute.manager [req-d904458d-45e0-46eb-813e-bf51f1ee8cc1 req-7f887e8f-6025-4d72-b250-c4c567f3b21a service nova] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Received unexpected event network-vif-plugged-9096db72-8476-4d1f-8fcc-cc5ddd54d61d for instance with vm_state building and task_state spawning. [ 859.223832] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.436s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.224350] env[61970]: DEBUG nova.compute.manager [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 859.226895] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.344s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.228417] env[61970]: INFO nova.compute.claims [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 859.240482] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Volume attach. Driver type: vmdk {{(pid=61970) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 859.240747] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288790', 'volume_id': 'f1e6410e-b97a-4846-b2db-fcd81bc4e2de', 'name': 'volume-f1e6410e-b97a-4846-b2db-fcd81bc4e2de', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b19cb516-b163-4bed-ba5b-139a0a18fc05', 'attached_at': '', 'detached_at': '', 'volume_id': 'f1e6410e-b97a-4846-b2db-fcd81bc4e2de', 'serial': 'f1e6410e-b97a-4846-b2db-fcd81bc4e2de'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 859.241632] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371fb0d0-87b3-4a3f-a72c-2d50725042cb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.266961] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f9a9038-29e7-43ae-886f-ee0cd4923c4b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.295863] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] volume-f1e6410e-b97a-4846-b2db-fcd81bc4e2de/volume-f1e6410e-b97a-4846-b2db-fcd81bc4e2de.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.297495] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b450bf3-dcb4-4450-95f2-c597c634c005 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.325863] env[61970]: DEBUG oslo_vmware.api [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 859.325863] env[61970]: value = "task-1355665" [ 859.325863] env[61970]: _type = "Task" [ 859.325863] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.334745] env[61970]: DEBUG oslo_vmware.api [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355665, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.372737] env[61970]: DEBUG nova.network.neutron [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Successfully updated port: 9096db72-8476-4d1f-8fcc-cc5ddd54d61d {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 859.558086] env[61970]: DEBUG oslo_vmware.api [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355663, 'name': PowerOffVM_Task, 'duration_secs': 0.203437} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.558382] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 859.558554] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 859.558818] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd6e272a-2434-43e4-b4f8-c816949f7a70 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.563465] env[61970]: DEBUG nova.network.neutron [req-6ee272ff-027d-4b0a-b581-9f317723e515 req-16d65c6c-81cf-4ee8-b4ed-cc3b6d42893b service nova] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Updated VIF entry in instance network info cache for port 1dfb626c-a7e0-455e-bffd-3eabada94ae1. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 859.563884] env[61970]: DEBUG nova.network.neutron [req-6ee272ff-027d-4b0a-b581-9f317723e515 req-16d65c6c-81cf-4ee8-b4ed-cc3b6d42893b service nova] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Updating instance_info_cache with network_info: [{"id": "1dfb626c-a7e0-455e-bffd-3eabada94ae1", "address": "fa:16:3e:b9:26:72", "network": {"id": "f08966ed-3f5b-4be7-bcb0-a0fe277e5367", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1451995592-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.180", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0429db7acc24d2eabeaad9efa9e66ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "650f7968-4522-4ba5-8304-1b9949951ed7", "external-id": "nsx-vlan-transportzone-568", "segmentation_id": 568, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1dfb626c-a7", "ovs_interfaceid": "1dfb626c-a7e0-455e-bffd-3eabada94ae1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.633166] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 859.633394] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 859.633593] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleting the datastore file [datastore2] d09dbde2-5daa-4ae9-9fd5-bfbb03fba136 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 859.634247] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7077fa1e-af98-4a6f-b4f6-45e7c660b693 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.639406] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355664, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.511526} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.640064] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] e1415314-f137-4ee3-a065-3d875ef9e7ff/e1415314-f137-4ee3-a065-3d875ef9e7ff.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 859.640243] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.640522] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-57cee8b5-9fd4-41da-86e9-77f75f9333fc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.643789] env[61970]: DEBUG oslo_vmware.api [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 859.643789] env[61970]: value = "task-1355667" [ 859.643789] env[61970]: _type = "Task" [ 859.643789] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.648610] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 859.648610] env[61970]: value = "task-1355668" [ 859.648610] env[61970]: _type = "Task" [ 859.648610] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.656196] env[61970]: DEBUG oslo_vmware.api [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355667, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.659136] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355668, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.735956] env[61970]: DEBUG nova.compute.utils [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 859.737587] env[61970]: DEBUG nova.compute.manager [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 859.737587] env[61970]: DEBUG nova.network.neutron [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 859.789349] env[61970]: DEBUG nova.policy [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '13aacb75013d408980ba5189c6071447', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c5b87b23cfe747379a4f42e6879f0f25', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 859.836970] env[61970]: DEBUG oslo_vmware.api [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355665, 'name': ReconfigVM_Task, 'duration_secs': 0.406996} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.836970] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Reconfigured VM instance instance-00000040 to attach disk [datastore1] volume-f1e6410e-b97a-4846-b2db-fcd81bc4e2de/volume-f1e6410e-b97a-4846-b2db-fcd81bc4e2de.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 859.840728] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-123f0dbf-2560-4575-89cc-9c93975b6c61 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.855232] env[61970]: DEBUG oslo_vmware.api [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 859.855232] env[61970]: value = "task-1355669" [ 859.855232] env[61970]: _type = "Task" [ 859.855232] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.862950] env[61970]: DEBUG oslo_vmware.api [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355669, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.878670] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Acquiring lock "refresh_cache-035fb797-21c1-4af9-8e66-deee8e9d083a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.878774] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Acquired lock "refresh_cache-035fb797-21c1-4af9-8e66-deee8e9d083a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.878955] env[61970]: DEBUG nova.network.neutron [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 860.061467] env[61970]: DEBUG nova.network.neutron [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Successfully created port: fbc5f4e5-3299-4e59-90aa-badfe6a2ca4c {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 860.067077] env[61970]: DEBUG oslo_concurrency.lockutils [req-6ee272ff-027d-4b0a-b581-9f317723e515 req-16d65c6c-81cf-4ee8-b4ed-cc3b6d42893b service nova] Releasing lock "refresh_cache-6838c6a4-8241-4b6f-ad58-3acfd7afdbf4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.156661] env[61970]: DEBUG oslo_vmware.api [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355667, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.161557] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355668, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062774} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.161824] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 860.162716] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6572c6d0-2695-46ce-9732-099bc413cb69 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.185905] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] e1415314-f137-4ee3-a065-3d875ef9e7ff/e1415314-f137-4ee3-a065-3d875ef9e7ff.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.186279] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f606fc2-0767-4d1f-b643-8b75ad400bd1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.206574] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 860.206574] env[61970]: value = "task-1355670" [ 860.206574] env[61970]: _type = "Task" [ 860.206574] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.215889] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355670, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.241644] env[61970]: DEBUG nova.compute.manager [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 860.368945] env[61970]: DEBUG oslo_vmware.api [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355669, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.423978] env[61970]: DEBUG nova.network.neutron [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 860.490014] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf0f2555-2d96-4048-8568-681b36190e0f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.497759] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14dcc1c2-4530-435e-9cc7-57cd5c2c51f3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.528892] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-099d1363-c17f-4d34-ae23-1081afe0d764 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.536132] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fafb5a77-ea41-4502-b717-427677d9e753 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.549519] env[61970]: DEBUG nova.compute.provider_tree [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 860.609219] env[61970]: DEBUG nova.network.neutron [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Updating instance_info_cache with network_info: [{"id": "9096db72-8476-4d1f-8fcc-cc5ddd54d61d", "address": "fa:16:3e:63:90:96", "network": {"id": "f27abf99-a724-47c4-aba5-984a7a69a03b", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1975217859-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1458b73cd8a4356816756e4c7d05925", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6db039c-542c-4544-a57d-ddcc6c1e8e45", "external-id": "nsx-vlan-transportzone-810", "segmentation_id": 810, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9096db72-84", "ovs_interfaceid": "9096db72-8476-4d1f-8fcc-cc5ddd54d61d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.656622] env[61970]: DEBUG oslo_vmware.api [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355667, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.528747} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.656889] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 860.657075] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 860.657253] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 860.657423] env[61970]: INFO nova.compute.manager [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Took 1.63 seconds to destroy the instance on the hypervisor. [ 860.657666] env[61970]: DEBUG oslo.service.loopingcall [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.657852] env[61970]: DEBUG nova.compute.manager [-] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 860.657946] env[61970]: DEBUG nova.network.neutron [-] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 860.716958] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355670, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.865537] env[61970]: DEBUG oslo_vmware.api [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355669, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.083214] env[61970]: DEBUG nova.scheduler.client.report [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Updated inventory for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with generation 85 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 861.083485] env[61970]: DEBUG nova.compute.provider_tree [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Updating resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa generation from 85 to 86 during operation: update_inventory {{(pid=61970) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 861.083667] env[61970]: DEBUG nova.compute.provider_tree [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 861.111799] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Releasing lock "refresh_cache-035fb797-21c1-4af9-8e66-deee8e9d083a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.112140] env[61970]: DEBUG nova.compute.manager [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Instance network_info: |[{"id": "9096db72-8476-4d1f-8fcc-cc5ddd54d61d", "address": "fa:16:3e:63:90:96", "network": {"id": "f27abf99-a724-47c4-aba5-984a7a69a03b", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1975217859-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1458b73cd8a4356816756e4c7d05925", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6db039c-542c-4544-a57d-ddcc6c1e8e45", "external-id": "nsx-vlan-transportzone-810", "segmentation_id": 810, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9096db72-84", "ovs_interfaceid": "9096db72-8476-4d1f-8fcc-cc5ddd54d61d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 861.112635] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:90:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e6db039c-542c-4544-a57d-ddcc6c1e8e45', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9096db72-8476-4d1f-8fcc-cc5ddd54d61d', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 861.120067] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Creating folder: Project (b1458b73cd8a4356816756e4c7d05925). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 861.120576] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b5c5647d-5334-4acd-b708-d60247811f98 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.131307] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Created folder: Project (b1458b73cd8a4356816756e4c7d05925) in parent group-v288740. [ 861.131470] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Creating folder: Instances. Parent ref: group-v288794. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 861.131720] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e3e0744e-aeff-4456-826a-d9d3854dca93 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.141269] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Created folder: Instances in parent group-v288794. [ 861.141497] env[61970]: DEBUG oslo.service.loopingcall [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.141676] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 861.141872] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-16d9c545-a79d-4b45-8c0b-a315f015f47e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.162588] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 861.162588] env[61970]: value = "task-1355673" [ 861.162588] env[61970]: _type = "Task" [ 861.162588] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.170626] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355673, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.217821] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355670, 'name': ReconfigVM_Task, 'duration_secs': 0.833777} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.218129] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Reconfigured VM instance instance-00000045 to attach disk [datastore2] e1415314-f137-4ee3-a065-3d875ef9e7ff/e1415314-f137-4ee3-a065-3d875ef9e7ff.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.218759] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-654321d6-2f2a-4d74-b98e-88a74a12cc76 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.225317] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 861.225317] env[61970]: value = "task-1355674" [ 861.225317] env[61970]: _type = "Task" [ 861.225317] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.238473] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355674, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.242931] env[61970]: DEBUG nova.compute.manager [req-e711daa0-c8f1-4557-a168-e167726a1b25 req-c575eb49-0796-4b65-9528-85f2f9eed5ee service nova] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Received event network-changed-9096db72-8476-4d1f-8fcc-cc5ddd54d61d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 861.243231] env[61970]: DEBUG nova.compute.manager [req-e711daa0-c8f1-4557-a168-e167726a1b25 req-c575eb49-0796-4b65-9528-85f2f9eed5ee service nova] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Refreshing instance network info cache due to event network-changed-9096db72-8476-4d1f-8fcc-cc5ddd54d61d. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 861.243377] env[61970]: DEBUG oslo_concurrency.lockutils [req-e711daa0-c8f1-4557-a168-e167726a1b25 req-c575eb49-0796-4b65-9528-85f2f9eed5ee service nova] Acquiring lock "refresh_cache-035fb797-21c1-4af9-8e66-deee8e9d083a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.243514] env[61970]: DEBUG oslo_concurrency.lockutils [req-e711daa0-c8f1-4557-a168-e167726a1b25 req-c575eb49-0796-4b65-9528-85f2f9eed5ee service nova] Acquired lock "refresh_cache-035fb797-21c1-4af9-8e66-deee8e9d083a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.243673] env[61970]: DEBUG nova.network.neutron [req-e711daa0-c8f1-4557-a168-e167726a1b25 req-c575eb49-0796-4b65-9528-85f2f9eed5ee service nova] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Refreshing network info cache for port 9096db72-8476-4d1f-8fcc-cc5ddd54d61d {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 861.252619] env[61970]: DEBUG nova.compute.manager [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 861.277806] env[61970]: DEBUG nova.virt.hardware [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 861.278063] env[61970]: DEBUG nova.virt.hardware [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 861.278224] env[61970]: DEBUG nova.virt.hardware [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 861.278402] env[61970]: DEBUG nova.virt.hardware [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 861.278545] env[61970]: DEBUG nova.virt.hardware [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 861.278690] env[61970]: DEBUG nova.virt.hardware [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 861.278895] env[61970]: DEBUG nova.virt.hardware [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 861.279068] env[61970]: DEBUG nova.virt.hardware [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 861.279239] env[61970]: DEBUG nova.virt.hardware [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 861.279400] env[61970]: DEBUG nova.virt.hardware [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 861.279581] env[61970]: DEBUG nova.virt.hardware [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 861.280796] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17064668-ad4a-4d6c-a057-b4b6943c3792 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.289033] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb743b0-ea84-4304-9ef9-a0a5529bcbca {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.368854] env[61970]: DEBUG oslo_vmware.api [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355669, 'name': ReconfigVM_Task, 'duration_secs': 1.014531} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.369257] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288790', 'volume_id': 'f1e6410e-b97a-4846-b2db-fcd81bc4e2de', 'name': 'volume-f1e6410e-b97a-4846-b2db-fcd81bc4e2de', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b19cb516-b163-4bed-ba5b-139a0a18fc05', 'attached_at': '', 'detached_at': '', 'volume_id': 'f1e6410e-b97a-4846-b2db-fcd81bc4e2de', 'serial': 'f1e6410e-b97a-4846-b2db-fcd81bc4e2de'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 861.438653] env[61970]: DEBUG nova.network.neutron [-] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.522628] env[61970]: DEBUG nova.compute.manager [req-56f4a4b8-6fd6-49c1-b592-d0c4d96b56a8 req-07726a41-d982-414c-ab9d-33ab7da59173 service nova] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Received event network-vif-plugged-fbc5f4e5-3299-4e59-90aa-badfe6a2ca4c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 861.522886] env[61970]: DEBUG oslo_concurrency.lockutils [req-56f4a4b8-6fd6-49c1-b592-d0c4d96b56a8 req-07726a41-d982-414c-ab9d-33ab7da59173 service nova] Acquiring lock "2d280f75-ed84-4db3-9067-8ccc719b0d2e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.523116] env[61970]: DEBUG oslo_concurrency.lockutils [req-56f4a4b8-6fd6-49c1-b592-d0c4d96b56a8 req-07726a41-d982-414c-ab9d-33ab7da59173 service nova] Lock "2d280f75-ed84-4db3-9067-8ccc719b0d2e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.523280] env[61970]: DEBUG oslo_concurrency.lockutils [req-56f4a4b8-6fd6-49c1-b592-d0c4d96b56a8 req-07726a41-d982-414c-ab9d-33ab7da59173 service nova] Lock "2d280f75-ed84-4db3-9067-8ccc719b0d2e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.523445] env[61970]: DEBUG nova.compute.manager [req-56f4a4b8-6fd6-49c1-b592-d0c4d96b56a8 req-07726a41-d982-414c-ab9d-33ab7da59173 service nova] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] No waiting events found dispatching network-vif-plugged-fbc5f4e5-3299-4e59-90aa-badfe6a2ca4c {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 861.523606] env[61970]: WARNING nova.compute.manager [req-56f4a4b8-6fd6-49c1-b592-d0c4d96b56a8 req-07726a41-d982-414c-ab9d-33ab7da59173 service nova] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Received unexpected event network-vif-plugged-fbc5f4e5-3299-4e59-90aa-badfe6a2ca4c for instance with vm_state building and task_state spawning. [ 861.588412] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.361s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.588926] env[61970]: DEBUG nova.compute.manager [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 861.592301] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.520s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.594047] env[61970]: INFO nova.compute.claims [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 861.672592] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355673, 'name': CreateVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.736876] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355674, 'name': Rename_Task, 'duration_secs': 0.129433} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.737180] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 861.737429] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d0b3a733-6d0a-4c82-a90e-0cdc87350ac2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.742941] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 861.742941] env[61970]: value = "task-1355675" [ 861.742941] env[61970]: _type = "Task" [ 861.742941] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.752179] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355675, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.941645] env[61970]: INFO nova.compute.manager [-] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Took 1.28 seconds to deallocate network for instance. [ 861.950247] env[61970]: DEBUG nova.network.neutron [req-e711daa0-c8f1-4557-a168-e167726a1b25 req-c575eb49-0796-4b65-9528-85f2f9eed5ee service nova] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Updated VIF entry in instance network info cache for port 9096db72-8476-4d1f-8fcc-cc5ddd54d61d. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 861.950247] env[61970]: DEBUG nova.network.neutron [req-e711daa0-c8f1-4557-a168-e167726a1b25 req-c575eb49-0796-4b65-9528-85f2f9eed5ee service nova] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Updating instance_info_cache with network_info: [{"id": "9096db72-8476-4d1f-8fcc-cc5ddd54d61d", "address": "fa:16:3e:63:90:96", "network": {"id": "f27abf99-a724-47c4-aba5-984a7a69a03b", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1975217859-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1458b73cd8a4356816756e4c7d05925", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6db039c-542c-4544-a57d-ddcc6c1e8e45", "external-id": "nsx-vlan-transportzone-810", "segmentation_id": 810, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9096db72-84", "ovs_interfaceid": "9096db72-8476-4d1f-8fcc-cc5ddd54d61d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.098155] env[61970]: DEBUG nova.compute.utils [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 862.101778] env[61970]: DEBUG nova.compute.manager [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 862.101948] env[61970]: DEBUG nova.network.neutron [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 862.131573] env[61970]: DEBUG nova.network.neutron [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Successfully updated port: fbc5f4e5-3299-4e59-90aa-badfe6a2ca4c {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 862.147300] env[61970]: DEBUG nova.policy [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '67a4e38db9894750b16edcf7f188220f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '19dae2316adb4dc980bd972c8c40c034', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 862.154829] env[61970]: DEBUG nova.compute.manager [req-fbaf8915-b584-4e9e-ad9b-8dd3970b236f req-6754c6bf-f568-4b59-abb2-1c9d736a04a6 service nova] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Received event network-changed-fbc5f4e5-3299-4e59-90aa-badfe6a2ca4c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.155039] env[61970]: DEBUG nova.compute.manager [req-fbaf8915-b584-4e9e-ad9b-8dd3970b236f req-6754c6bf-f568-4b59-abb2-1c9d736a04a6 service nova] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Refreshing instance network info cache due to event network-changed-fbc5f4e5-3299-4e59-90aa-badfe6a2ca4c. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 862.155254] env[61970]: DEBUG oslo_concurrency.lockutils [req-fbaf8915-b584-4e9e-ad9b-8dd3970b236f req-6754c6bf-f568-4b59-abb2-1c9d736a04a6 service nova] Acquiring lock "refresh_cache-2d280f75-ed84-4db3-9067-8ccc719b0d2e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.155408] env[61970]: DEBUG oslo_concurrency.lockutils [req-fbaf8915-b584-4e9e-ad9b-8dd3970b236f req-6754c6bf-f568-4b59-abb2-1c9d736a04a6 service nova] Acquired lock "refresh_cache-2d280f75-ed84-4db3-9067-8ccc719b0d2e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.155590] env[61970]: DEBUG nova.network.neutron [req-fbaf8915-b584-4e9e-ad9b-8dd3970b236f req-6754c6bf-f568-4b59-abb2-1c9d736a04a6 service nova] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Refreshing network info cache for port fbc5f4e5-3299-4e59-90aa-badfe6a2ca4c {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 862.175194] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355673, 'name': CreateVM_Task, 'duration_secs': 0.547551} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.175417] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 862.176280] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.176753] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.176894] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 862.179426] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d409c3c-45ed-4f27-ba99-6f9a4c45bb96 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.184855] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Waiting for the task: (returnval){ [ 862.184855] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5236d701-c7e4-8abd-f795-4a99807a9610" [ 862.184855] env[61970]: _type = "Task" [ 862.184855] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.195012] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5236d701-c7e4-8abd-f795-4a99807a9610, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.253047] env[61970]: DEBUG oslo_vmware.api [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355675, 'name': PowerOnVM_Task, 'duration_secs': 0.438482} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.253357] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 862.253534] env[61970]: INFO nova.compute.manager [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Took 5.86 seconds to spawn the instance on the hypervisor. [ 862.253711] env[61970]: DEBUG nova.compute.manager [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 862.254474] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ba2cd0-8f4c-4b18-851c-5dda207a1cd1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.408445] env[61970]: DEBUG nova.objects.instance [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lazy-loading 'flavor' on Instance uuid b19cb516-b163-4bed-ba5b-139a0a18fc05 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.440703] env[61970]: DEBUG nova.network.neutron [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Successfully created port: 7bfd07fd-97bc-42b5-9571-0807d864f9bc {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 862.449591] env[61970]: DEBUG oslo_concurrency.lockutils [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.452544] env[61970]: DEBUG oslo_concurrency.lockutils [req-e711daa0-c8f1-4557-a168-e167726a1b25 req-c575eb49-0796-4b65-9528-85f2f9eed5ee service nova] Releasing lock "refresh_cache-035fb797-21c1-4af9-8e66-deee8e9d083a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.452860] env[61970]: DEBUG nova.compute.manager [req-e711daa0-c8f1-4557-a168-e167726a1b25 req-c575eb49-0796-4b65-9528-85f2f9eed5ee service nova] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Received event network-vif-deleted-201e52ca-273b-4df3-b32a-d9dba51cb5c7 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.452968] env[61970]: INFO nova.compute.manager [req-e711daa0-c8f1-4557-a168-e167726a1b25 req-c575eb49-0796-4b65-9528-85f2f9eed5ee service nova] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Neutron deleted interface 201e52ca-273b-4df3-b32a-d9dba51cb5c7; detaching it from the instance and deleting it from the info cache [ 862.453127] env[61970]: DEBUG nova.network.neutron [req-e711daa0-c8f1-4557-a168-e167726a1b25 req-c575eb49-0796-4b65-9528-85f2f9eed5ee service nova] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.607161] env[61970]: DEBUG nova.compute.manager [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 862.637191] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "refresh_cache-2d280f75-ed84-4db3-9067-8ccc719b0d2e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.696098] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5236d701-c7e4-8abd-f795-4a99807a9610, 'name': SearchDatastore_Task, 'duration_secs': 0.010951} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.696440] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.696812] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 862.697012] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.697262] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.697433] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 862.697731] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6fe3f9cb-5094-48e0-a0d7-6c90c7fd9285 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.702981] env[61970]: DEBUG nova.network.neutron [req-fbaf8915-b584-4e9e-ad9b-8dd3970b236f req-6754c6bf-f568-4b59-abb2-1c9d736a04a6 service nova] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 862.717332] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 862.717513] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 862.718261] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3766de6f-bc33-471d-ada2-5c71cdc5b85c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.726664] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Waiting for the task: (returnval){ [ 862.726664] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b167f7-4a8f-37cc-0a31-264a20c68d08" [ 862.726664] env[61970]: _type = "Task" [ 862.726664] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.734946] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b167f7-4a8f-37cc-0a31-264a20c68d08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.774993] env[61970]: INFO nova.compute.manager [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Took 23.55 seconds to build instance. [ 862.799965] env[61970]: DEBUG nova.network.neutron [req-fbaf8915-b584-4e9e-ad9b-8dd3970b236f req-6754c6bf-f568-4b59-abb2-1c9d736a04a6 service nova] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.885766] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-596b498c-9156-4590-9dff-89f1df7f7b94 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.893332] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c202658-88df-4278-964d-fc28c22743c7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.928503] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f9e01e-0e22-4196-97f8-5cfe7106e62b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.931378] env[61970]: DEBUG oslo_concurrency.lockutils [None req-17f9cc05-dec7-4390-a4c2-0367cc1f4c3c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "b19cb516-b163-4bed-ba5b-139a0a18fc05" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.338s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.937230] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03dc3345-922b-49a2-83c2-944d80385c19 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.951093] env[61970]: DEBUG nova.compute.provider_tree [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.955223] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4f8242fb-616f-46ec-90aa-b6ff3a70c5a0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.965642] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-845c5973-3f0a-4506-a98a-f26fa7b7b5ee {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.992462] env[61970]: DEBUG nova.compute.manager [req-e711daa0-c8f1-4557-a168-e167726a1b25 req-c575eb49-0796-4b65-9528-85f2f9eed5ee service nova] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Detach interface failed, port_id=201e52ca-273b-4df3-b32a-d9dba51cb5c7, reason: Instance d09dbde2-5daa-4ae9-9fd5-bfbb03fba136 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 863.076923] env[61970]: DEBUG oslo_concurrency.lockutils [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "b19cb516-b163-4bed-ba5b-139a0a18fc05" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.077293] env[61970]: DEBUG oslo_concurrency.lockutils [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "b19cb516-b163-4bed-ba5b-139a0a18fc05" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.077521] env[61970]: DEBUG oslo_concurrency.lockutils [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "b19cb516-b163-4bed-ba5b-139a0a18fc05-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.077709] env[61970]: DEBUG oslo_concurrency.lockutils [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "b19cb516-b163-4bed-ba5b-139a0a18fc05-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.077924] env[61970]: DEBUG oslo_concurrency.lockutils [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "b19cb516-b163-4bed-ba5b-139a0a18fc05-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.079917] env[61970]: INFO nova.compute.manager [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Terminating instance [ 863.082826] env[61970]: DEBUG nova.compute.manager [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 863.082826] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 863.082826] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-62a7fb9b-166e-4205-8157-d43b43df9255 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.089703] env[61970]: DEBUG oslo_vmware.api [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 863.089703] env[61970]: value = "task-1355676" [ 863.089703] env[61970]: _type = "Task" [ 863.089703] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.098458] env[61970]: DEBUG oslo_vmware.api [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355676, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.237486] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b167f7-4a8f-37cc-0a31-264a20c68d08, 'name': SearchDatastore_Task, 'duration_secs': 0.008662} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.238497] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2dec534a-4574-4f17-b25f-e7cb4835fbe4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.244704] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Waiting for the task: (returnval){ [ 863.244704] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e417c9-cf4d-4099-787b-95eb2123b55e" [ 863.244704] env[61970]: _type = "Task" [ 863.244704] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.253130] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e417c9-cf4d-4099-787b-95eb2123b55e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.276490] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6df7edb-a7bf-4b3d-a559-a714d3d34abe tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Lock "e1415314-f137-4ee3-a065-3d875ef9e7ff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.958s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.303042] env[61970]: DEBUG oslo_concurrency.lockutils [req-fbaf8915-b584-4e9e-ad9b-8dd3970b236f req-6754c6bf-f568-4b59-abb2-1c9d736a04a6 service nova] Releasing lock "refresh_cache-2d280f75-ed84-4db3-9067-8ccc719b0d2e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.303042] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquired lock "refresh_cache-2d280f75-ed84-4db3-9067-8ccc719b0d2e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.303042] env[61970]: DEBUG nova.network.neutron [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 863.453510] env[61970]: DEBUG nova.scheduler.client.report [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.599491] env[61970]: DEBUG oslo_vmware.api [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355676, 'name': PowerOffVM_Task, 'duration_secs': 0.256542} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.599755] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 863.599952] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Volume detach. Driver type: vmdk {{(pid=61970) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 863.600153] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288790', 'volume_id': 'f1e6410e-b97a-4846-b2db-fcd81bc4e2de', 'name': 'volume-f1e6410e-b97a-4846-b2db-fcd81bc4e2de', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b19cb516-b163-4bed-ba5b-139a0a18fc05', 'attached_at': '', 'detached_at': '', 'volume_id': 'f1e6410e-b97a-4846-b2db-fcd81bc4e2de', 'serial': 'f1e6410e-b97a-4846-b2db-fcd81bc4e2de'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 863.600871] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f01f7c2-2f2b-4ba5-bd35-ad9e77b55969 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.622139] env[61970]: DEBUG nova.compute.manager [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 863.624596] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b0f1d9-7229-4505-8300-b0c87f88f601 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.631490] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe912ef5-5a36-4b87-824a-6fe42b257a6f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.651662] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc5b666-a9eb-422c-a284-ce0f038c094d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.656674] env[61970]: DEBUG nova.virt.hardware [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 863.656901] env[61970]: DEBUG nova.virt.hardware [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 863.657135] env[61970]: DEBUG nova.virt.hardware [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 863.657382] env[61970]: DEBUG nova.virt.hardware [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 863.657539] env[61970]: DEBUG nova.virt.hardware [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 863.657739] env[61970]: DEBUG nova.virt.hardware [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 863.657901] env[61970]: DEBUG nova.virt.hardware [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 863.658072] env[61970]: DEBUG nova.virt.hardware [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 863.658243] env[61970]: DEBUG nova.virt.hardware [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 863.658402] env[61970]: DEBUG nova.virt.hardware [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 863.658569] env[61970]: DEBUG nova.virt.hardware [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 863.659282] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e8d0da3-f3a3-49a1-bed2-c0191541fcd8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.671538] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] The volume has not been displaced from its original location: [datastore1] volume-f1e6410e-b97a-4846-b2db-fcd81bc4e2de/volume-f1e6410e-b97a-4846-b2db-fcd81bc4e2de.vmdk. No consolidation needed. {{(pid=61970) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 863.678300] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Reconfiguring VM instance instance-00000040 to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 863.678888] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0111fd4a-ae1b-4c93-b62e-8b8254dcd7c5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.694572] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-660d76f9-b44f-4dde-9fb4-86c914b88946 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.699445] env[61970]: DEBUG oslo_vmware.api [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 863.699445] env[61970]: value = "task-1355677" [ 863.699445] env[61970]: _type = "Task" [ 863.699445] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.715813] env[61970]: DEBUG oslo_vmware.api [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355677, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.754914] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e417c9-cf4d-4099-787b-95eb2123b55e, 'name': SearchDatastore_Task, 'duration_secs': 0.010284} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.755213] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.755474] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 035fb797-21c1-4af9-8e66-deee8e9d083a/035fb797-21c1-4af9-8e66-deee8e9d083a.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 863.755756] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0f111bff-21e2-4dee-9448-6e55d6e66342 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.761611] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Waiting for the task: (returnval){ [ 863.761611] env[61970]: value = "task-1355678" [ 863.761611] env[61970]: _type = "Task" [ 863.761611] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.770530] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': task-1355678, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.815854] env[61970]: INFO nova.compute.manager [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Rebuilding instance [ 863.850025] env[61970]: DEBUG nova.network.neutron [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 863.857418] env[61970]: DEBUG nova.compute.manager [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 863.858330] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-206e503e-81f3-4e1d-b678-77287b6af096 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.958843] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.366s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.959494] env[61970]: DEBUG nova.compute.manager [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 863.962408] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.035s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.964628] env[61970]: INFO nova.compute.claims [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 864.021622] env[61970]: DEBUG nova.network.neutron [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Updating instance_info_cache with network_info: [{"id": "fbc5f4e5-3299-4e59-90aa-badfe6a2ca4c", "address": "fa:16:3e:0c:2c:40", "network": {"id": "40524b52-3115-4853-b278-29feded80350", "bridge": "br-int", "label": "tempest-ImagesTestJSON-729386347-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c5b87b23cfe747379a4f42e6879f0f25", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbc5f4e5-32", "ovs_interfaceid": "fbc5f4e5-3299-4e59-90aa-badfe6a2ca4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.039368] env[61970]: DEBUG nova.compute.manager [req-d8fe96fd-a715-4d4b-8fbe-620a0118c1ae req-0e798580-cce7-47c5-9a4e-23195295ed2b service nova] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Received event network-vif-plugged-7bfd07fd-97bc-42b5-9571-0807d864f9bc {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.039636] env[61970]: DEBUG oslo_concurrency.lockutils [req-d8fe96fd-a715-4d4b-8fbe-620a0118c1ae req-0e798580-cce7-47c5-9a4e-23195295ed2b service nova] Acquiring lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.039796] env[61970]: DEBUG oslo_concurrency.lockutils [req-d8fe96fd-a715-4d4b-8fbe-620a0118c1ae req-0e798580-cce7-47c5-9a4e-23195295ed2b service nova] Lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.040443] env[61970]: DEBUG oslo_concurrency.lockutils [req-d8fe96fd-a715-4d4b-8fbe-620a0118c1ae req-0e798580-cce7-47c5-9a4e-23195295ed2b service nova] Lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.040629] env[61970]: DEBUG nova.compute.manager [req-d8fe96fd-a715-4d4b-8fbe-620a0118c1ae req-0e798580-cce7-47c5-9a4e-23195295ed2b service nova] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] No waiting events found dispatching network-vif-plugged-7bfd07fd-97bc-42b5-9571-0807d864f9bc {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 864.040795] env[61970]: WARNING nova.compute.manager [req-d8fe96fd-a715-4d4b-8fbe-620a0118c1ae req-0e798580-cce7-47c5-9a4e-23195295ed2b service nova] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Received unexpected event network-vif-plugged-7bfd07fd-97bc-42b5-9571-0807d864f9bc for instance with vm_state building and task_state spawning. [ 864.160386] env[61970]: DEBUG nova.network.neutron [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Successfully updated port: 7bfd07fd-97bc-42b5-9571-0807d864f9bc {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 864.210528] env[61970]: DEBUG oslo_vmware.api [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355677, 'name': ReconfigVM_Task, 'duration_secs': 0.244036} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.210912] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Reconfigured VM instance instance-00000040 to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 864.216818] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2be66f87-bad5-4d35-af38-bbb2ca3552e3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.233771] env[61970]: DEBUG oslo_vmware.api [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 864.233771] env[61970]: value = "task-1355679" [ 864.233771] env[61970]: _type = "Task" [ 864.233771] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.242933] env[61970]: DEBUG oslo_vmware.api [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355679, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.271409] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': task-1355678, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.369368] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 864.369704] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d372ad4d-3a7b-4da9-95b3-f8448efe80af {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.376693] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 864.376693] env[61970]: value = "task-1355680" [ 864.376693] env[61970]: _type = "Task" [ 864.376693] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.384814] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355680, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.469204] env[61970]: DEBUG nova.compute.utils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 864.471069] env[61970]: DEBUG nova.compute.manager [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 864.471223] env[61970]: DEBUG nova.network.neutron [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 864.514201] env[61970]: DEBUG nova.policy [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b07eb2474b1c4fbb98207a4d435500a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd9c85ee0fccc4373b4898f0256a69d38', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 864.526587] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Releasing lock "refresh_cache-2d280f75-ed84-4db3-9067-8ccc719b0d2e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.526902] env[61970]: DEBUG nova.compute.manager [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Instance network_info: |[{"id": "fbc5f4e5-3299-4e59-90aa-badfe6a2ca4c", "address": "fa:16:3e:0c:2c:40", "network": {"id": "40524b52-3115-4853-b278-29feded80350", "bridge": "br-int", "label": "tempest-ImagesTestJSON-729386347-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c5b87b23cfe747379a4f42e6879f0f25", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbc5f4e5-32", "ovs_interfaceid": "fbc5f4e5-3299-4e59-90aa-badfe6a2ca4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 864.527377] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:2c:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ce058b2d-df85-481c-a996-cc179d534f1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fbc5f4e5-3299-4e59-90aa-badfe6a2ca4c', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 864.535203] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Creating folder: Project (c5b87b23cfe747379a4f42e6879f0f25). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 864.535480] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7f89fbb4-fe38-4181-a984-7a128b043c12 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.545823] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Created folder: Project (c5b87b23cfe747379a4f42e6879f0f25) in parent group-v288740. [ 864.546020] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Creating folder: Instances. Parent ref: group-v288797. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 864.546303] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f471621d-c99e-4a54-b555-b7f5e2097ef7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.560200] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Created folder: Instances in parent group-v288797. [ 864.560425] env[61970]: DEBUG oslo.service.loopingcall [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.560637] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 864.560870] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aa32387e-ea7d-47f3-827f-f4a6005868e0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.581999] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 864.581999] env[61970]: value = "task-1355683" [ 864.581999] env[61970]: _type = "Task" [ 864.581999] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.589115] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355683, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.664719] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "refresh_cache-15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.664913] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquired lock "refresh_cache-15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.665083] env[61970]: DEBUG nova.network.neutron [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 864.744297] env[61970]: DEBUG oslo_vmware.api [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355679, 'name': ReconfigVM_Task, 'duration_secs': 0.189948} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.744659] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288790', 'volume_id': 'f1e6410e-b97a-4846-b2db-fcd81bc4e2de', 'name': 'volume-f1e6410e-b97a-4846-b2db-fcd81bc4e2de', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b19cb516-b163-4bed-ba5b-139a0a18fc05', 'attached_at': '', 'detached_at': '', 'volume_id': 'f1e6410e-b97a-4846-b2db-fcd81bc4e2de', 'serial': 'f1e6410e-b97a-4846-b2db-fcd81bc4e2de'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 864.745145] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 864.745980] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63a98f0-268a-4409-bdaf-2b3f5cc672e7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.758316] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 864.758316] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e9920d57-c3d4-4d52-af1d-313733dfaf77 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.776999] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': task-1355678, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510486} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.777518] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 035fb797-21c1-4af9-8e66-deee8e9d083a/035fb797-21c1-4af9-8e66-deee8e9d083a.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 864.777626] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 864.777799] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-75f657e3-5d1d-4281-a0da-9f236be10918 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.782800] env[61970]: DEBUG nova.network.neutron [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Successfully created port: f0126a7e-adec-4042-8b0e-5eb42bbd1572 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 864.786206] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Waiting for the task: (returnval){ [ 864.786206] env[61970]: value = "task-1355685" [ 864.786206] env[61970]: _type = "Task" [ 864.786206] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.794126] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': task-1355685, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.861519] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 864.861662] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 864.861743] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Deleting the datastore file [datastore1] b19cb516-b163-4bed-ba5b-139a0a18fc05 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 864.862012] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae83e90e-a53d-4e2e-8572-0a1d4329b629 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.869149] env[61970]: DEBUG oslo_vmware.api [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 864.869149] env[61970]: value = "task-1355686" [ 864.869149] env[61970]: _type = "Task" [ 864.869149] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.876980] env[61970]: DEBUG oslo_vmware.api [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355686, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.884490] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355680, 'name': PowerOffVM_Task, 'duration_secs': 0.200279} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.884727] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 864.884953] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 864.885824] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c64853-f71f-4b15-bbe6-462786b3cb1a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.893182] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 864.893466] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-75c1c660-aaf3-4161-93c6-65b39a5af598 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.921413] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 864.921618] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 864.921797] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Deleting the datastore file [datastore2] e1415314-f137-4ee3-a065-3d875ef9e7ff {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 864.922085] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-07c34e1d-e0ae-4ae1-866a-bbd9099e1d1b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.928675] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 864.928675] env[61970]: value = "task-1355688" [ 864.928675] env[61970]: _type = "Task" [ 864.928675] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.937299] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355688, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.976298] env[61970]: DEBUG nova.compute.manager [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 865.094142] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355683, 'name': CreateVM_Task, 'duration_secs': 0.442226} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.094319] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 865.095014] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.095186] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.095593] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 865.095861] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aba3960e-4363-4144-970b-b22dc937c08b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.101897] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for the task: (returnval){ [ 865.101897] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ed15a0-6b5d-d1c3-a41c-a080757a9c22" [ 865.101897] env[61970]: _type = "Task" [ 865.101897] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.112038] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ed15a0-6b5d-d1c3-a41c-a080757a9c22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.214231] env[61970]: DEBUG nova.network.neutron [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 865.233329] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e48c4dc-0702-4ff9-a5f2-df4f4a7019fd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.241929] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ced5654-e70e-4891-aa1f-7b81f2ed2c7c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.276339] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3d85387-7744-4384-9fe5-f1ab6e1575e9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.284680] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d69fc5e-e1a0-4fb2-8d71-e6a7272c939b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.301468] env[61970]: DEBUG nova.compute.provider_tree [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 865.305566] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': task-1355685, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081983} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.306059] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 865.306807] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdbe17ff-b051-459b-92c1-8fae8d52cf54 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.329366] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 035fb797-21c1-4af9-8e66-deee8e9d083a/035fb797-21c1-4af9-8e66-deee8e9d083a.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.332616] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8307442-b44f-4165-a0db-f0c8965f7371 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.352504] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Waiting for the task: (returnval){ [ 865.352504] env[61970]: value = "task-1355689" [ 865.352504] env[61970]: _type = "Task" [ 865.352504] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.360885] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': task-1355689, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.380252] env[61970]: DEBUG oslo_vmware.api [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355686, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.439118] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355688, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.475374} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.439441] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 865.439636] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 865.439829] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 865.475649] env[61970]: DEBUG nova.network.neutron [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Updating instance_info_cache with network_info: [{"id": "7bfd07fd-97bc-42b5-9571-0807d864f9bc", "address": "fa:16:3e:89:e6:54", "network": {"id": "42def9ae-4eed-41e5-a538-81eaced2cdd4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2047751481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19dae2316adb4dc980bd972c8c40c034", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c1b8689-a9b4-4972-beb9-6a1c8de1dc88", "external-id": "nsx-vlan-transportzone-455", "segmentation_id": 455, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bfd07fd-97", "ovs_interfaceid": "7bfd07fd-97bc-42b5-9571-0807d864f9bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.615465] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ed15a0-6b5d-d1c3-a41c-a080757a9c22, 'name': SearchDatastore_Task, 'duration_secs': 0.065518} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.615465] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.615730] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.616031] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.616202] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.616381] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 865.616650] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dfcef78c-0b96-46b1-af71-da515822d801 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.624472] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 865.624674] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 865.625453] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ada12b07-c232-4118-a5c7-cdc7805659bd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.630559] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for the task: (returnval){ [ 865.630559] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5261acb1-3d06-2193-91b9-ae12f5aa9217" [ 865.630559] env[61970]: _type = "Task" [ 865.630559] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.637669] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5261acb1-3d06-2193-91b9-ae12f5aa9217, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.836436] env[61970]: DEBUG nova.scheduler.client.report [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Updated inventory for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with generation 86 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 865.836726] env[61970]: DEBUG nova.compute.provider_tree [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Updating resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa generation from 86 to 87 during operation: update_inventory {{(pid=61970) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 865.836947] env[61970]: DEBUG nova.compute.provider_tree [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 865.863949] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': task-1355689, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.879162] env[61970]: DEBUG oslo_vmware.api [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355686, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.568987} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.879399] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 865.879583] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 865.879758] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 865.879954] env[61970]: INFO nova.compute.manager [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Took 2.80 seconds to destroy the instance on the hypervisor. [ 865.880220] env[61970]: DEBUG oslo.service.loopingcall [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.880396] env[61970]: DEBUG nova.compute.manager [-] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 865.880535] env[61970]: DEBUG nova.network.neutron [-] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 865.977996] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Releasing lock "refresh_cache-15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.978551] env[61970]: DEBUG nova.compute.manager [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Instance network_info: |[{"id": "7bfd07fd-97bc-42b5-9571-0807d864f9bc", "address": "fa:16:3e:89:e6:54", "network": {"id": "42def9ae-4eed-41e5-a538-81eaced2cdd4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2047751481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19dae2316adb4dc980bd972c8c40c034", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c1b8689-a9b4-4972-beb9-6a1c8de1dc88", "external-id": "nsx-vlan-transportzone-455", "segmentation_id": 455, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bfd07fd-97", "ovs_interfaceid": "7bfd07fd-97bc-42b5-9571-0807d864f9bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 865.978730] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:e6:54', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9c1b8689-a9b4-4972-beb9-6a1c8de1dc88', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7bfd07fd-97bc-42b5-9571-0807d864f9bc', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 865.986980] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Creating folder: Project (19dae2316adb4dc980bd972c8c40c034). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 865.987710] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ad48fa5-c14e-4c75-bb70-c01923b85962 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.993230] env[61970]: DEBUG nova.compute.manager [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 865.998541] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Created folder: Project (19dae2316adb4dc980bd972c8c40c034) in parent group-v288740. [ 865.998729] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Creating folder: Instances. Parent ref: group-v288800. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 865.998966] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ab30c62d-bb9d-4da5-9d62-140aa5723d9e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.010434] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Created folder: Instances in parent group-v288800. [ 866.010679] env[61970]: DEBUG oslo.service.loopingcall [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 866.012539] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 866.014623] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8a5b0262-0d17-44e2-b03c-0491459934da {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.029290] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 866.029526] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 866.029681] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 866.029861] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 866.030014] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 866.030166] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 866.030374] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 866.030531] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 866.030696] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 866.030855] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 866.031038] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 866.031863] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c7e8146-3835-48bf-a1da-7c4cef88262f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.040661] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3062b9f8-c9d5-4440-9fb7-4ced8f9bb2a0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.044290] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 866.044290] env[61970]: value = "task-1355692" [ 866.044290] env[61970]: _type = "Task" [ 866.044290] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.060473] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355692, 'name': CreateVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.066092] env[61970]: DEBUG nova.compute.manager [req-b89b8403-de52-427b-b5fc-27ccafbd43ea req-37af2a45-70bf-42fe-8cfd-a63eac829925 service nova] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Received event network-changed-7bfd07fd-97bc-42b5-9571-0807d864f9bc {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.066171] env[61970]: DEBUG nova.compute.manager [req-b89b8403-de52-427b-b5fc-27ccafbd43ea req-37af2a45-70bf-42fe-8cfd-a63eac829925 service nova] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Refreshing instance network info cache due to event network-changed-7bfd07fd-97bc-42b5-9571-0807d864f9bc. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 866.066327] env[61970]: DEBUG oslo_concurrency.lockutils [req-b89b8403-de52-427b-b5fc-27ccafbd43ea req-37af2a45-70bf-42fe-8cfd-a63eac829925 service nova] Acquiring lock "refresh_cache-15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.066551] env[61970]: DEBUG oslo_concurrency.lockutils [req-b89b8403-de52-427b-b5fc-27ccafbd43ea req-37af2a45-70bf-42fe-8cfd-a63eac829925 service nova] Acquired lock "refresh_cache-15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.066642] env[61970]: DEBUG nova.network.neutron [req-b89b8403-de52-427b-b5fc-27ccafbd43ea req-37af2a45-70bf-42fe-8cfd-a63eac829925 service nova] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Refreshing network info cache for port 7bfd07fd-97bc-42b5-9571-0807d864f9bc {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 866.147428] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5261acb1-3d06-2193-91b9-ae12f5aa9217, 'name': SearchDatastore_Task, 'duration_secs': 0.01375} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.151539] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8041dcb8-094b-4d88-85a6-5fdbf02aa311 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.163048] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for the task: (returnval){ [ 866.163048] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f1817c-1dc4-bf43-4ab6-6cc6c151e3f2" [ 866.163048] env[61970]: _type = "Task" [ 866.163048] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.179162] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f1817c-1dc4-bf43-4ab6-6cc6c151e3f2, 'name': SearchDatastore_Task, 'duration_secs': 0.009137} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.179759] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.180244] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 2d280f75-ed84-4db3-9067-8ccc719b0d2e/2d280f75-ed84-4db3-9067-8ccc719b0d2e.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 866.180662] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9fc2c687-7e20-4f7b-b663-aa4f63ca7390 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.189885] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for the task: (returnval){ [ 866.189885] env[61970]: value = "task-1355693" [ 866.189885] env[61970]: _type = "Task" [ 866.189885] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.207811] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355693, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.249999] env[61970]: DEBUG nova.compute.manager [req-309de235-ac53-4c17-bc9e-16eb1a537038 req-f197c377-ff5a-4aa4-be5a-f3a8326ed093 service nova] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Received event network-vif-deleted-90ade7be-34cd-42db-b38c-0e64cba25ec4 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.250167] env[61970]: INFO nova.compute.manager [req-309de235-ac53-4c17-bc9e-16eb1a537038 req-f197c377-ff5a-4aa4-be5a-f3a8326ed093 service nova] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Neutron deleted interface 90ade7be-34cd-42db-b38c-0e64cba25ec4; detaching it from the instance and deleting it from the info cache [ 866.250448] env[61970]: DEBUG nova.network.neutron [req-309de235-ac53-4c17-bc9e-16eb1a537038 req-f197c377-ff5a-4aa4-be5a-f3a8326ed093 service nova] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.342648] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.343284] env[61970]: DEBUG nova.compute.manager [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 866.346431] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.984s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.348537] env[61970]: INFO nova.compute.claims [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 866.367176] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': task-1355689, 'name': ReconfigVM_Task, 'duration_secs': 0.569051} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.367176] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 035fb797-21c1-4af9-8e66-deee8e9d083a/035fb797-21c1-4af9-8e66-deee8e9d083a.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.367805] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ee30fe76-f92d-47b2-af3e-7f6690cd714e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.374884] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Waiting for the task: (returnval){ [ 866.374884] env[61970]: value = "task-1355694" [ 866.374884] env[61970]: _type = "Task" [ 866.374884] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.384327] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': task-1355694, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.480456] env[61970]: DEBUG nova.virt.hardware [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 866.480728] env[61970]: DEBUG nova.virt.hardware [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 866.480936] env[61970]: DEBUG nova.virt.hardware [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 866.481217] env[61970]: DEBUG nova.virt.hardware [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 866.481422] env[61970]: DEBUG nova.virt.hardware [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 866.481612] env[61970]: DEBUG nova.virt.hardware [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 866.481875] env[61970]: DEBUG nova.virt.hardware [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 866.482087] env[61970]: DEBUG nova.virt.hardware [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 866.482299] env[61970]: DEBUG nova.virt.hardware [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 866.482708] env[61970]: DEBUG nova.virt.hardware [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 866.482945] env[61970]: DEBUG nova.virt.hardware [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 866.484027] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6582fa-81e1-430b-85ec-61bb73b1e375 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.492923] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8ef977b-16f6-4780-9023-a21f88c71fa8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.508198] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Instance VIF info [] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 866.514559] env[61970]: DEBUG oslo.service.loopingcall [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 866.515361] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 866.515645] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-58a3ba81-cbba-4c1f-bc7b-01f59ac14515 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.534523] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 866.534523] env[61970]: value = "task-1355695" [ 866.534523] env[61970]: _type = "Task" [ 866.534523] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.545867] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355695, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.553990] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355692, 'name': CreateVM_Task, 'duration_secs': 0.339041} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.554180] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 866.554971] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.555192] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.555521] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 866.555914] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77c7a421-4797-4120-b799-a959450895e7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.561427] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 866.561427] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]520b10e7-29ed-4834-7521-52dbca713b48" [ 866.561427] env[61970]: _type = "Task" [ 866.561427] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.572527] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]520b10e7-29ed-4834-7521-52dbca713b48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.574599] env[61970]: DEBUG nova.network.neutron [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Successfully updated port: f0126a7e-adec-4042-8b0e-5eb42bbd1572 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 866.657992] env[61970]: DEBUG nova.network.neutron [-] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.701936] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355693, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48427} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.702138] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 2d280f75-ed84-4db3-9067-8ccc719b0d2e/2d280f75-ed84-4db3-9067-8ccc719b0d2e.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 866.702482] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.702884] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-be784573-0f15-4267-96a5-a3ab57d9bd4c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.711602] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for the task: (returnval){ [ 866.711602] env[61970]: value = "task-1355696" [ 866.711602] env[61970]: _type = "Task" [ 866.711602] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.719745] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355696, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.753900] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3eb9db25-f3ba-4cb2-b8ea-cb72c140d774 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.762832] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90e9821-3cb0-4be0-8470-5ab98225fd07 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.790065] env[61970]: DEBUG nova.compute.manager [req-309de235-ac53-4c17-bc9e-16eb1a537038 req-f197c377-ff5a-4aa4-be5a-f3a8326ed093 service nova] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Detach interface failed, port_id=90ade7be-34cd-42db-b38c-0e64cba25ec4, reason: Instance b19cb516-b163-4bed-ba5b-139a0a18fc05 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 866.828105] env[61970]: DEBUG nova.network.neutron [req-b89b8403-de52-427b-b5fc-27ccafbd43ea req-37af2a45-70bf-42fe-8cfd-a63eac829925 service nova] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Updated VIF entry in instance network info cache for port 7bfd07fd-97bc-42b5-9571-0807d864f9bc. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 866.828483] env[61970]: DEBUG nova.network.neutron [req-b89b8403-de52-427b-b5fc-27ccafbd43ea req-37af2a45-70bf-42fe-8cfd-a63eac829925 service nova] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Updating instance_info_cache with network_info: [{"id": "7bfd07fd-97bc-42b5-9571-0807d864f9bc", "address": "fa:16:3e:89:e6:54", "network": {"id": "42def9ae-4eed-41e5-a538-81eaced2cdd4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2047751481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19dae2316adb4dc980bd972c8c40c034", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c1b8689-a9b4-4972-beb9-6a1c8de1dc88", "external-id": "nsx-vlan-transportzone-455", "segmentation_id": 455, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bfd07fd-97", "ovs_interfaceid": "7bfd07fd-97bc-42b5-9571-0807d864f9bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.854188] env[61970]: DEBUG nova.compute.utils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 866.857724] env[61970]: DEBUG nova.compute.manager [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 866.857901] env[61970]: DEBUG nova.network.neutron [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 866.887044] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': task-1355694, 'name': Rename_Task, 'duration_secs': 0.367122} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.887044] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 866.887044] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e29ea80a-d529-4409-88fb-661c4d1c6940 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.894044] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Waiting for the task: (returnval){ [ 866.894044] env[61970]: value = "task-1355697" [ 866.894044] env[61970]: _type = "Task" [ 866.894044] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.898208] env[61970]: DEBUG nova.policy [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b07eb2474b1c4fbb98207a4d435500a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd9c85ee0fccc4373b4898f0256a69d38', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 866.905064] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': task-1355697, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.044621] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355695, 'name': CreateVM_Task, 'duration_secs': 0.326987} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.044893] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 867.045352] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.072569] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]520b10e7-29ed-4834-7521-52dbca713b48, 'name': SearchDatastore_Task, 'duration_secs': 0.053663} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.072982] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.073308] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 867.073602] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.073810] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.074058] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 867.074414] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.074771] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 867.075105] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e37ef34b-ff1f-4cab-b2fa-ff31db2742e1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.077249] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63938d8d-451a-45fb-95ed-62d02936fc28 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.079125] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "refresh_cache-99200b46-7eb7-4ca2-b352-b91b4ad076af" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.079319] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquired lock "refresh_cache-99200b46-7eb7-4ca2-b352-b91b4ad076af" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.079515] env[61970]: DEBUG nova.network.neutron [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 867.086827] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 867.086827] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525498b0-0d4d-b6e9-0f84-9793dfebceec" [ 867.086827] env[61970]: _type = "Task" [ 867.086827] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.090838] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 867.090838] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 867.092722] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5514d95f-f113-4ee6-885e-e56c1bc44938 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.099102] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525498b0-0d4d-b6e9-0f84-9793dfebceec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.102586] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 867.102586] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5227c13c-1b2b-de06-1a2f-e5cb01e8592f" [ 867.102586] env[61970]: _type = "Task" [ 867.102586] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.111009] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5227c13c-1b2b-de06-1a2f-e5cb01e8592f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.164926] env[61970]: INFO nova.compute.manager [-] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Took 1.28 seconds to deallocate network for instance. [ 867.199978] env[61970]: DEBUG nova.network.neutron [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Successfully created port: 3cec35ca-1e33-4861-8a90-71573ae58d5c {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 867.224159] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355696, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060726} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.224447] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 867.225280] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08190cbf-d461-48bd-b410-594d37d814e3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.249552] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 2d280f75-ed84-4db3-9067-8ccc719b0d2e/2d280f75-ed84-4db3-9067-8ccc719b0d2e.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.249860] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df17510d-11fd-4b3f-9670-038a95a23d04 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.271539] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for the task: (returnval){ [ 867.271539] env[61970]: value = "task-1355698" [ 867.271539] env[61970]: _type = "Task" [ 867.271539] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.280811] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355698, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.331245] env[61970]: DEBUG oslo_concurrency.lockutils [req-b89b8403-de52-427b-b5fc-27ccafbd43ea req-37af2a45-70bf-42fe-8cfd-a63eac829925 service nova] Releasing lock "refresh_cache-15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.358420] env[61970]: DEBUG nova.compute.manager [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 867.411644] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': task-1355697, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.597494] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525498b0-0d4d-b6e9-0f84-9793dfebceec, 'name': SearchDatastore_Task, 'duration_secs': 0.048456} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.600054] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.600164] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 867.600421] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.613968] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5227c13c-1b2b-de06-1a2f-e5cb01e8592f, 'name': SearchDatastore_Task, 'duration_secs': 0.054358} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.616491] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16faa87d-5ebb-4710-bad1-cec59f2b2e88 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.620581] env[61970]: DEBUG nova.network.neutron [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 867.622769] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4eb653cf-e1e3-4a87-b234-429b496e66d5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.631499] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2f18aee-2009-4f09-a18e-021b7f201f56 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.634657] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 867.634657] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a54af0-4299-1e8d-1cfc-d40ef25c2fbf" [ 867.634657] env[61970]: _type = "Task" [ 867.634657] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.666016] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63645c97-5b5f-48aa-8843-41e7e81d930f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.672723] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a54af0-4299-1e8d-1cfc-d40ef25c2fbf, 'name': SearchDatastore_Task, 'duration_secs': 0.030395} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.672858] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.673133] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa/15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 867.673413] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.673598] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 867.673800] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-220b72bf-5948-4b93-97a5-8124a531619f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.677448] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-24270748-edba-45fa-8f52-4b7f2b9013f7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.680884] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8393ca0-3e39-4152-bcf3-e9795e624514 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.687285] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 867.687285] env[61970]: value = "task-1355699" [ 867.687285] env[61970]: _type = "Task" [ 867.687285] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.698638] env[61970]: DEBUG nova.compute.provider_tree [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.703287] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 867.703287] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 867.704450] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7281f36f-d8fb-4823-936e-8d065c08d3fe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.710359] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355699, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.713328] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 867.713328] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52c0876a-11a0-c2c7-5eaa-789eccfc33fd" [ 867.713328] env[61970]: _type = "Task" [ 867.713328] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.722404] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52c0876a-11a0-c2c7-5eaa-789eccfc33fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.723370] env[61970]: INFO nova.compute.manager [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Took 0.56 seconds to detach 1 volumes for instance. [ 867.785199] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355698, 'name': ReconfigVM_Task, 'duration_secs': 0.313293} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.785364] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 2d280f75-ed84-4db3-9067-8ccc719b0d2e/2d280f75-ed84-4db3-9067-8ccc719b0d2e.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.786118] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1d469c63-63fa-4ac9-bc38-23c4c8e010e2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.793373] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for the task: (returnval){ [ 867.793373] env[61970]: value = "task-1355700" [ 867.793373] env[61970]: _type = "Task" [ 867.793373] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.801303] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355700, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.824024] env[61970]: DEBUG nova.network.neutron [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Updating instance_info_cache with network_info: [{"id": "f0126a7e-adec-4042-8b0e-5eb42bbd1572", "address": "fa:16:3e:44:0f:97", "network": {"id": "a394e9eb-7879-4d19-b05f-358c85dfbe48", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-767180013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d9c85ee0fccc4373b4898f0256a69d38", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0126a7e-ad", "ovs_interfaceid": "f0126a7e-adec-4042-8b0e-5eb42bbd1572", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.913470] env[61970]: DEBUG oslo_vmware.api [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': task-1355697, 'name': PowerOnVM_Task, 'duration_secs': 0.783938} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.913827] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 867.914441] env[61970]: INFO nova.compute.manager [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Took 9.10 seconds to spawn the instance on the hypervisor. [ 867.914441] env[61970]: DEBUG nova.compute.manager [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.915259] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cc7e57a-4dbe-4fc4-a2bc-7f528e61bcbf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.092370] env[61970]: DEBUG nova.compute.manager [req-26036894-bcd7-4122-b495-5a21d8b65811 req-7dbc5d70-5855-46f5-a7d9-79ad66440e17 service nova] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Received event network-vif-plugged-f0126a7e-adec-4042-8b0e-5eb42bbd1572 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 868.092644] env[61970]: DEBUG oslo_concurrency.lockutils [req-26036894-bcd7-4122-b495-5a21d8b65811 req-7dbc5d70-5855-46f5-a7d9-79ad66440e17 service nova] Acquiring lock "99200b46-7eb7-4ca2-b352-b91b4ad076af-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.092804] env[61970]: DEBUG oslo_concurrency.lockutils [req-26036894-bcd7-4122-b495-5a21d8b65811 req-7dbc5d70-5855-46f5-a7d9-79ad66440e17 service nova] Lock "99200b46-7eb7-4ca2-b352-b91b4ad076af-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.092989] env[61970]: DEBUG oslo_concurrency.lockutils [req-26036894-bcd7-4122-b495-5a21d8b65811 req-7dbc5d70-5855-46f5-a7d9-79ad66440e17 service nova] Lock "99200b46-7eb7-4ca2-b352-b91b4ad076af-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.093155] env[61970]: DEBUG nova.compute.manager [req-26036894-bcd7-4122-b495-5a21d8b65811 req-7dbc5d70-5855-46f5-a7d9-79ad66440e17 service nova] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] No waiting events found dispatching network-vif-plugged-f0126a7e-adec-4042-8b0e-5eb42bbd1572 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 868.093307] env[61970]: WARNING nova.compute.manager [req-26036894-bcd7-4122-b495-5a21d8b65811 req-7dbc5d70-5855-46f5-a7d9-79ad66440e17 service nova] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Received unexpected event network-vif-plugged-f0126a7e-adec-4042-8b0e-5eb42bbd1572 for instance with vm_state building and task_state spawning. [ 868.093649] env[61970]: DEBUG nova.compute.manager [req-26036894-bcd7-4122-b495-5a21d8b65811 req-7dbc5d70-5855-46f5-a7d9-79ad66440e17 service nova] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Received event network-changed-f0126a7e-adec-4042-8b0e-5eb42bbd1572 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 868.093852] env[61970]: DEBUG nova.compute.manager [req-26036894-bcd7-4122-b495-5a21d8b65811 req-7dbc5d70-5855-46f5-a7d9-79ad66440e17 service nova] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Refreshing instance network info cache due to event network-changed-f0126a7e-adec-4042-8b0e-5eb42bbd1572. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 868.094023] env[61970]: DEBUG oslo_concurrency.lockutils [req-26036894-bcd7-4122-b495-5a21d8b65811 req-7dbc5d70-5855-46f5-a7d9-79ad66440e17 service nova] Acquiring lock "refresh_cache-99200b46-7eb7-4ca2-b352-b91b4ad076af" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.201058] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355699, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.208767] env[61970]: DEBUG nova.scheduler.client.report [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 868.226669] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52c0876a-11a0-c2c7-5eaa-789eccfc33fd, 'name': SearchDatastore_Task, 'duration_secs': 0.03321} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.227376] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12aae6c9-9d93-4c78-a40c-85ba88e901cb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.232666] env[61970]: DEBUG oslo_concurrency.lockutils [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.234573] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 868.234573] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52915fe9-1aac-80ee-fcda-a448b63344b1" [ 868.234573] env[61970]: _type = "Task" [ 868.234573] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.244586] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52915fe9-1aac-80ee-fcda-a448b63344b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.303739] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355700, 'name': Rename_Task, 'duration_secs': 0.132745} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.304105] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 868.304247] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9ad47a8d-d442-449f-875e-987bdb9611aa {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.312015] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for the task: (returnval){ [ 868.312015] env[61970]: value = "task-1355701" [ 868.312015] env[61970]: _type = "Task" [ 868.312015] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.321156] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355701, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.326990] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Releasing lock "refresh_cache-99200b46-7eb7-4ca2-b352-b91b4ad076af" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.327418] env[61970]: DEBUG nova.compute.manager [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Instance network_info: |[{"id": "f0126a7e-adec-4042-8b0e-5eb42bbd1572", "address": "fa:16:3e:44:0f:97", "network": {"id": "a394e9eb-7879-4d19-b05f-358c85dfbe48", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-767180013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d9c85ee0fccc4373b4898f0256a69d38", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0126a7e-ad", "ovs_interfaceid": "f0126a7e-adec-4042-8b0e-5eb42bbd1572", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 868.327777] env[61970]: DEBUG oslo_concurrency.lockutils [req-26036894-bcd7-4122-b495-5a21d8b65811 req-7dbc5d70-5855-46f5-a7d9-79ad66440e17 service nova] Acquired lock "refresh_cache-99200b46-7eb7-4ca2-b352-b91b4ad076af" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.327973] env[61970]: DEBUG nova.network.neutron [req-26036894-bcd7-4122-b495-5a21d8b65811 req-7dbc5d70-5855-46f5-a7d9-79ad66440e17 service nova] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Refreshing network info cache for port f0126a7e-adec-4042-8b0e-5eb42bbd1572 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 868.329428] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:0f:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c24464bb-bb6b-43a2-bdcd-8086ad1a307f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f0126a7e-adec-4042-8b0e-5eb42bbd1572', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 868.338348] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Creating folder: Project (d9c85ee0fccc4373b4898f0256a69d38). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 868.339536] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-111662a1-14f7-4ee2-a7cf-7034bcfc491a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.353674] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Created folder: Project (d9c85ee0fccc4373b4898f0256a69d38) in parent group-v288740. [ 868.353885] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Creating folder: Instances. Parent ref: group-v288804. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 868.354129] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d47c6651-9f65-4e9a-8fd3-ba1962ea53da {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.367021] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Created folder: Instances in parent group-v288804. [ 868.367448] env[61970]: DEBUG oslo.service.loopingcall [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 868.367675] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 868.367904] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ef7743cf-2ba6-4402-ba2c-2ad8f86236e4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.384225] env[61970]: DEBUG nova.compute.manager [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 868.392081] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 868.392081] env[61970]: value = "task-1355704" [ 868.392081] env[61970]: _type = "Task" [ 868.392081] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.402940] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355704, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.411155] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 868.411402] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 868.411558] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 868.411746] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 868.411906] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 868.412053] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 868.412290] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 868.412443] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 868.412636] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 868.412817] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 868.412993] env[61970]: DEBUG nova.virt.hardware [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 868.413864] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaec618f-9bdc-43ab-be58-f86df87a6f91 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.422054] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a08c3f-f9bd-421e-8c18-6f01f0361ab0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.443244] env[61970]: INFO nova.compute.manager [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Took 28.62 seconds to build instance. [ 868.700238] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355699, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.747501} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.700533] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa/15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 868.700824] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 868.701121] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-496c2866-b80e-4a5b-a494-646034f3f9bf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.712018] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 868.712018] env[61970]: value = "task-1355705" [ 868.712018] env[61970]: _type = "Task" [ 868.712018] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.716338] env[61970]: DEBUG nova.compute.manager [req-9c07d7ce-4c1b-4603-bf23-419a65fc2258 req-ba108ad3-a039-4692-9d60-ad646dba5752 service nova] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Received event network-vif-plugged-3cec35ca-1e33-4861-8a90-71573ae58d5c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 868.716633] env[61970]: DEBUG oslo_concurrency.lockutils [req-9c07d7ce-4c1b-4603-bf23-419a65fc2258 req-ba108ad3-a039-4692-9d60-ad646dba5752 service nova] Acquiring lock "e2185ed0-2bc1-4718-b47d-57150b5e60ba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.716937] env[61970]: DEBUG oslo_concurrency.lockutils [req-9c07d7ce-4c1b-4603-bf23-419a65fc2258 req-ba108ad3-a039-4692-9d60-ad646dba5752 service nova] Lock "e2185ed0-2bc1-4718-b47d-57150b5e60ba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.717654] env[61970]: DEBUG oslo_concurrency.lockutils [req-9c07d7ce-4c1b-4603-bf23-419a65fc2258 req-ba108ad3-a039-4692-9d60-ad646dba5752 service nova] Lock "e2185ed0-2bc1-4718-b47d-57150b5e60ba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.717654] env[61970]: DEBUG nova.compute.manager [req-9c07d7ce-4c1b-4603-bf23-419a65fc2258 req-ba108ad3-a039-4692-9d60-ad646dba5752 service nova] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] No waiting events found dispatching network-vif-plugged-3cec35ca-1e33-4861-8a90-71573ae58d5c {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 868.717654] env[61970]: WARNING nova.compute.manager [req-9c07d7ce-4c1b-4603-bf23-419a65fc2258 req-ba108ad3-a039-4692-9d60-ad646dba5752 service nova] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Received unexpected event network-vif-plugged-3cec35ca-1e33-4861-8a90-71573ae58d5c for instance with vm_state building and task_state spawning. [ 868.718929] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.719413] env[61970]: DEBUG nova.compute.manager [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 868.727052] env[61970]: DEBUG oslo_concurrency.lockutils [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.859s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.728264] env[61970]: INFO nova.compute.claims [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 868.730952] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355705, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.747740] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52915fe9-1aac-80ee-fcda-a448b63344b1, 'name': SearchDatastore_Task, 'duration_secs': 0.049553} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.748007] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.748297] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] e1415314-f137-4ee3-a065-3d875ef9e7ff/e1415314-f137-4ee3-a065-3d875ef9e7ff.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 868.748572] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2cfe9a7b-96f2-4006-97c0-9998dfed48d5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.755698] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 868.755698] env[61970]: value = "task-1355706" [ 868.755698] env[61970]: _type = "Task" [ 868.755698] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.764362] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355706, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.822167] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355701, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.832009] env[61970]: DEBUG nova.network.neutron [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Successfully updated port: 3cec35ca-1e33-4861-8a90-71573ae58d5c {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 868.902409] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355704, 'name': CreateVM_Task, 'duration_secs': 0.333877} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.902639] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 868.903827] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.903827] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.903972] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 868.904178] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8012d5fb-af20-481f-bdc3-7e1b4efb8fb0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.910826] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 868.910826] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d41081-e5e8-aafc-7ddb-cc4e46d7e689" [ 868.910826] env[61970]: _type = "Task" [ 868.910826] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.921606] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d41081-e5e8-aafc-7ddb-cc4e46d7e689, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.945437] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1511940b-1a69-4899-a15d-8e224c9054a3 tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Lock "035fb797-21c1-4af9-8e66-deee8e9d083a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.010s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.219572] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355705, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.487163} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.219862] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 869.220650] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7f3b4b-9777-44fc-952a-7657704cb1cc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.235595] env[61970]: DEBUG nova.compute.utils [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 869.247597] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa/15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 869.248730] env[61970]: DEBUG nova.network.neutron [req-26036894-bcd7-4122-b495-5a21d8b65811 req-7dbc5d70-5855-46f5-a7d9-79ad66440e17 service nova] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Updated VIF entry in instance network info cache for port f0126a7e-adec-4042-8b0e-5eb42bbd1572. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 869.249056] env[61970]: DEBUG nova.network.neutron [req-26036894-bcd7-4122-b495-5a21d8b65811 req-7dbc5d70-5855-46f5-a7d9-79ad66440e17 service nova] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Updating instance_info_cache with network_info: [{"id": "f0126a7e-adec-4042-8b0e-5eb42bbd1572", "address": "fa:16:3e:44:0f:97", "network": {"id": "a394e9eb-7879-4d19-b05f-358c85dfbe48", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-767180013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d9c85ee0fccc4373b4898f0256a69d38", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0126a7e-ad", "ovs_interfaceid": "f0126a7e-adec-4042-8b0e-5eb42bbd1572", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.250259] env[61970]: DEBUG nova.compute.manager [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 869.250430] env[61970]: DEBUG nova.network.neutron [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 869.252171] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80803a83-4b93-4a26-8c4d-a8698a31588d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.268314] env[61970]: DEBUG oslo_concurrency.lockutils [req-26036894-bcd7-4122-b495-5a21d8b65811 req-7dbc5d70-5855-46f5-a7d9-79ad66440e17 service nova] Releasing lock "refresh_cache-99200b46-7eb7-4ca2-b352-b91b4ad076af" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.277337] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355706, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.278569] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 869.278569] env[61970]: value = "task-1355707" [ 869.278569] env[61970]: _type = "Task" [ 869.278569] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.286461] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355707, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.323690] env[61970]: DEBUG oslo_vmware.api [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355701, 'name': PowerOnVM_Task, 'duration_secs': 0.555877} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.323916] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 869.324147] env[61970]: INFO nova.compute.manager [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Took 8.07 seconds to spawn the instance on the hypervisor. [ 869.324287] env[61970]: DEBUG nova.compute.manager [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.325082] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31a9cd5-6d48-49c7-8c9d-c2a08e5e9eb9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.333886] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "refresh_cache-e2185ed0-2bc1-4718-b47d-57150b5e60ba" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.334014] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquired lock "refresh_cache-e2185ed0-2bc1-4718-b47d-57150b5e60ba" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.334301] env[61970]: DEBUG nova.network.neutron [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 869.347755] env[61970]: DEBUG nova.policy [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b45dc9d71c0e4693ba42322efde6c572', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5c64cb9232fb413cbd7627dcf077e9ef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 869.422174] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d41081-e5e8-aafc-7ddb-cc4e46d7e689, 'name': SearchDatastore_Task, 'duration_secs': 0.04158} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.422560] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.423018] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 869.423347] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.423508] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.423696] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 869.423989] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5423f237-611e-41c9-9e52-31796a5172cd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.439849] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 869.440901] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 869.441795] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ee6e112-5246-453d-be6a-e9721d4d6dc5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.450386] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 869.450386] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b5f88b-e988-bb59-8c50-cc3f7e8ffc32" [ 869.450386] env[61970]: _type = "Task" [ 869.450386] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.463749] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b5f88b-e988-bb59-8c50-cc3f7e8ffc32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.544342] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Acquiring lock "035fb797-21c1-4af9-8e66-deee8e9d083a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.544595] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Lock "035fb797-21c1-4af9-8e66-deee8e9d083a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.544805] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Acquiring lock "035fb797-21c1-4af9-8e66-deee8e9d083a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.545097] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Lock "035fb797-21c1-4af9-8e66-deee8e9d083a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.545307] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Lock "035fb797-21c1-4af9-8e66-deee8e9d083a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.549111] env[61970]: INFO nova.compute.manager [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Terminating instance [ 869.551320] env[61970]: DEBUG nova.compute.manager [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 869.551548] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 869.552509] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d6f4ac3-31e8-4514-bb5e-930c0cdad7b7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.561506] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 869.561861] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bb94e348-04ad-4933-9245-d167a98ab4db {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.569272] env[61970]: DEBUG oslo_vmware.api [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Waiting for the task: (returnval){ [ 869.569272] env[61970]: value = "task-1355708" [ 869.569272] env[61970]: _type = "Task" [ 869.569272] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.578067] env[61970]: DEBUG oslo_vmware.api [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': task-1355708, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.726857] env[61970]: DEBUG nova.network.neutron [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Successfully created port: 32711b6a-a9cc-472d-a634-e3b60711bffe {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 869.748843] env[61970]: DEBUG nova.compute.manager [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 869.779082] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355706, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.962069} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.779436] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] e1415314-f137-4ee3-a065-3d875ef9e7ff/e1415314-f137-4ee3-a065-3d875ef9e7ff.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 869.779599] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 869.785437] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-08a58864-4ba2-46b7-8b7f-d14b89b204d1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.793474] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355707, 'name': ReconfigVM_Task, 'duration_secs': 0.510453} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.794832] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa/15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 869.795631] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 869.795631] env[61970]: value = "task-1355709" [ 869.795631] env[61970]: _type = "Task" [ 869.795631] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.795865] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-904a3753-8493-4436-8b1f-5087da169eb1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.812639] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 869.812639] env[61970]: value = "task-1355710" [ 869.812639] env[61970]: _type = "Task" [ 869.812639] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.815549] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355709, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.825205] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355710, 'name': Rename_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.846507] env[61970]: INFO nova.compute.manager [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Took 26.21 seconds to build instance. [ 869.897696] env[61970]: DEBUG nova.network.neutron [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 869.960509] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b5f88b-e988-bb59-8c50-cc3f7e8ffc32, 'name': SearchDatastore_Task, 'duration_secs': 0.050431} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.963690] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22957d40-9541-4279-8017-ad652e552279 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.968711] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 869.968711] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a79ce1-971a-955e-dedf-1520c047dc68" [ 869.968711] env[61970]: _type = "Task" [ 869.968711] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.976582] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a79ce1-971a-955e-dedf-1520c047dc68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.045338] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b1b5a85-239d-484b-b8b1-3f1f2a1b83d9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.053151] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31414581-229f-4143-b223-cc9228208c5d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.092831] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1039eb5-2aa6-4efc-8904-21d63db9ead6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.099520] env[61970]: DEBUG oslo_vmware.api [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': task-1355708, 'name': PowerOffVM_Task, 'duration_secs': 0.295331} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.102778] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 870.102778] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 870.102778] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bcd24621-e23d-40ae-9c17-bc25709c7259 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.105361] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d218ba95-01bc-472c-b175-8954a010f940 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.119900] env[61970]: DEBUG nova.compute.provider_tree [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.148599] env[61970]: DEBUG nova.network.neutron [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Updating instance_info_cache with network_info: [{"id": "3cec35ca-1e33-4861-8a90-71573ae58d5c", "address": "fa:16:3e:5c:88:8d", "network": {"id": "a394e9eb-7879-4d19-b05f-358c85dfbe48", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-767180013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d9c85ee0fccc4373b4898f0256a69d38", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cec35ca-1e", "ovs_interfaceid": "3cec35ca-1e33-4861-8a90-71573ae58d5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.172761] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 870.172990] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 870.173192] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Deleting the datastore file [datastore2] 035fb797-21c1-4af9-8e66-deee8e9d083a {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 870.173748] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8958b45c-b1e6-4cf5-a043-12519d6aa0d8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.180326] env[61970]: DEBUG oslo_vmware.api [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Waiting for the task: (returnval){ [ 870.180326] env[61970]: value = "task-1355712" [ 870.180326] env[61970]: _type = "Task" [ 870.180326] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.189385] env[61970]: DEBUG oslo_vmware.api [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': task-1355712, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.308464] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355709, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.104345} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.308736] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 870.309509] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc3912c-57b7-40c0-bd65-d351b344fdbe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.329604] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] e1415314-f137-4ee3-a065-3d875ef9e7ff/e1415314-f137-4ee3-a065-3d875ef9e7ff.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.332872] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bbaf7045-2b10-4999-94f1-a8a53fbd40e1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.348960] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d1c340eb-1fe4-4d9f-83a7-4fecc7182d45 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "2d280f75-ed84-4db3-9067-8ccc719b0d2e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.868s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.353993] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355710, 'name': Rename_Task, 'duration_secs': 0.149332} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.355232] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 870.355583] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 870.355583] env[61970]: value = "task-1355713" [ 870.355583] env[61970]: _type = "Task" [ 870.355583] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.355753] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef74ccb5-da26-48f7-a57d-14a0f055efa8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.365487] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355713, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.366607] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 870.366607] env[61970]: value = "task-1355714" [ 870.366607] env[61970]: _type = "Task" [ 870.366607] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.479775] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a79ce1-971a-955e-dedf-1520c047dc68, 'name': SearchDatastore_Task, 'duration_secs': 0.008738} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.480069] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.480861] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 99200b46-7eb7-4ca2-b352-b91b4ad076af/99200b46-7eb7-4ca2-b352-b91b4ad076af.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 870.480861] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-afb29dfa-3ee7-4657-a257-99fbceba245a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.486776] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 870.486776] env[61970]: value = "task-1355715" [ 870.486776] env[61970]: _type = "Task" [ 870.486776] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.494756] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355715, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.627399] env[61970]: DEBUG nova.scheduler.client.report [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 870.653018] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Releasing lock "refresh_cache-e2185ed0-2bc1-4718-b47d-57150b5e60ba" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.653018] env[61970]: DEBUG nova.compute.manager [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Instance network_info: |[{"id": "3cec35ca-1e33-4861-8a90-71573ae58d5c", "address": "fa:16:3e:5c:88:8d", "network": {"id": "a394e9eb-7879-4d19-b05f-358c85dfbe48", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-767180013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d9c85ee0fccc4373b4898f0256a69d38", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cec35ca-1e", "ovs_interfaceid": "3cec35ca-1e33-4861-8a90-71573ae58d5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 870.653361] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:88:8d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c24464bb-bb6b-43a2-bdcd-8086ad1a307f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3cec35ca-1e33-4861-8a90-71573ae58d5c', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 870.658731] env[61970]: DEBUG oslo.service.loopingcall [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.659279] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 870.659508] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4a05ab07-68c5-4cb1-8a57-4881b4c99b52 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.686108] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 870.686108] env[61970]: value = "task-1355716" [ 870.686108] env[61970]: _type = "Task" [ 870.686108] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.693715] env[61970]: DEBUG oslo_vmware.api [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Task: {'id': task-1355712, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15147} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.694653] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 870.694855] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 870.695124] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 870.695370] env[61970]: INFO nova.compute.manager [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Took 1.14 seconds to destroy the instance on the hypervisor. [ 870.695634] env[61970]: DEBUG oslo.service.loopingcall [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.699300] env[61970]: DEBUG nova.compute.manager [-] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 870.699462] env[61970]: DEBUG nova.network.neutron [-] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 870.701296] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355716, 'name': CreateVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.742713] env[61970]: DEBUG nova.compute.manager [req-49addb6f-99e7-4f39-8a62-422cec556d52 req-1d186010-fb88-47e0-bd1a-7365e2be4e4a service nova] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Received event network-changed-3cec35ca-1e33-4861-8a90-71573ae58d5c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.743037] env[61970]: DEBUG nova.compute.manager [req-49addb6f-99e7-4f39-8a62-422cec556d52 req-1d186010-fb88-47e0-bd1a-7365e2be4e4a service nova] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Refreshing instance network info cache due to event network-changed-3cec35ca-1e33-4861-8a90-71573ae58d5c. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 870.743332] env[61970]: DEBUG oslo_concurrency.lockutils [req-49addb6f-99e7-4f39-8a62-422cec556d52 req-1d186010-fb88-47e0-bd1a-7365e2be4e4a service nova] Acquiring lock "refresh_cache-e2185ed0-2bc1-4718-b47d-57150b5e60ba" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.743606] env[61970]: DEBUG oslo_concurrency.lockutils [req-49addb6f-99e7-4f39-8a62-422cec556d52 req-1d186010-fb88-47e0-bd1a-7365e2be4e4a service nova] Acquired lock "refresh_cache-e2185ed0-2bc1-4718-b47d-57150b5e60ba" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.743879] env[61970]: DEBUG nova.network.neutron [req-49addb6f-99e7-4f39-8a62-422cec556d52 req-1d186010-fb88-47e0-bd1a-7365e2be4e4a service nova] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Refreshing network info cache for port 3cec35ca-1e33-4861-8a90-71573ae58d5c {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 870.760906] env[61970]: DEBUG nova.compute.manager [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 870.788999] env[61970]: DEBUG nova.virt.hardware [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.789337] env[61970]: DEBUG nova.virt.hardware [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.789525] env[61970]: DEBUG nova.virt.hardware [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.789713] env[61970]: DEBUG nova.virt.hardware [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.789860] env[61970]: DEBUG nova.virt.hardware [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.790018] env[61970]: DEBUG nova.virt.hardware [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.790307] env[61970]: DEBUG nova.virt.hardware [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.790530] env[61970]: DEBUG nova.virt.hardware [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.790707] env[61970]: DEBUG nova.virt.hardware [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.790888] env[61970]: DEBUG nova.virt.hardware [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.791155] env[61970]: DEBUG nova.virt.hardware [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.792748] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc80a20-55f3-405c-a969-7405b4802c13 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.802632] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c15cbc2-e003-42dd-97b8-f4ef71576265 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.867780] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355713, 'name': ReconfigVM_Task, 'duration_secs': 0.462645} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.874801] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Reconfigured VM instance instance-00000045 to attach disk [datastore2] e1415314-f137-4ee3-a065-3d875ef9e7ff/e1415314-f137-4ee3-a065-3d875ef9e7ff.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 870.875666] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8eecb146-98b2-4250-9769-65335105ad3e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.884040] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355714, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.885807] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 870.885807] env[61970]: value = "task-1355717" [ 870.885807] env[61970]: _type = "Task" [ 870.885807] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.895416] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355717, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.977756] env[61970]: DEBUG nova.compute.manager [req-346edeeb-9b6c-43f8-9b4c-53c0926223b4 req-a3b718e8-1eb2-44a4-8500-d1eed9b010f9 service nova] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Received event network-vif-deleted-9096db72-8476-4d1f-8fcc-cc5ddd54d61d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.977756] env[61970]: INFO nova.compute.manager [req-346edeeb-9b6c-43f8-9b4c-53c0926223b4 req-a3b718e8-1eb2-44a4-8500-d1eed9b010f9 service nova] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Neutron deleted interface 9096db72-8476-4d1f-8fcc-cc5ddd54d61d; detaching it from the instance and deleting it from the info cache [ 870.977756] env[61970]: DEBUG nova.network.neutron [req-346edeeb-9b6c-43f8-9b4c-53c0926223b4 req-a3b718e8-1eb2-44a4-8500-d1eed9b010f9 service nova] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.997948] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355715, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476816} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.998307] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 99200b46-7eb7-4ca2-b352-b91b4ad076af/99200b46-7eb7-4ca2-b352-b91b4ad076af.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 870.998778] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 870.998872] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a6007ddb-256f-4212-91cc-a9371ed34fb7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.005942] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 871.005942] env[61970]: value = "task-1355718" [ 871.005942] env[61970]: _type = "Task" [ 871.005942] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.015724] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355718, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.132500] env[61970]: DEBUG oslo_concurrency.lockutils [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.406s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.133301] env[61970]: DEBUG nova.compute.manager [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 871.136163] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 13.245s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.136384] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.136597] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61970) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 871.136949] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.984s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.137325] env[61970]: DEBUG nova.objects.instance [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lazy-loading 'resources' on Instance uuid 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.142199] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41698803-cedf-478f-b48f-a49be084a0f3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.148640] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec0beec-40fb-46da-887e-bdcf74bb42f0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.164162] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b84fe829-d843-46e7-9da7-6c572ce0df3a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.171408] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6cdf81b-3376-4167-9b5e-2bcfb51f5af1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.205747] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181130MB free_disk=150GB free_vcpus=48 pci_devices=None {{(pid=61970) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 871.205893] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.214708] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355716, 'name': CreateVM_Task, 'duration_secs': 0.506523} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.214869] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 871.215578] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.215764] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.216120] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 871.216382] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3823c058-b270-4cf9-822d-5ebf480bd878 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.220959] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 871.220959] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5264904e-cf5b-cd1f-02fd-4b4bf01f0770" [ 871.220959] env[61970]: _type = "Task" [ 871.220959] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.228775] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5264904e-cf5b-cd1f-02fd-4b4bf01f0770, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.377416] env[61970]: DEBUG oslo_vmware.api [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355714, 'name': PowerOnVM_Task, 'duration_secs': 0.684929} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.377687] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 871.377888] env[61970]: INFO nova.compute.manager [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Took 7.76 seconds to spawn the instance on the hypervisor. [ 871.378077] env[61970]: DEBUG nova.compute.manager [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 871.378866] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11825c21-b69e-43e8-a440-f56a629b5e53 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.394880] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355717, 'name': Rename_Task, 'duration_secs': 0.229063} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.395151] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 871.395419] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-290c250f-a837-48e5-bcc8-6d6e5934b52c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.401458] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 871.401458] env[61970]: value = "task-1355719" [ 871.401458] env[61970]: _type = "Task" [ 871.401458] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.409554] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355719, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.410501] env[61970]: DEBUG nova.network.neutron [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Successfully updated port: 32711b6a-a9cc-472d-a634-e3b60711bffe {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 871.451059] env[61970]: DEBUG nova.compute.manager [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 871.452083] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb0724a-729c-4a0d-a3f5-21f837faba65 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.454833] env[61970]: DEBUG nova.network.neutron [-] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.479978] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-16d5c789-6420-4f23-9ec4-1445abfebafd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.490038] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9463d20c-1a8d-4e04-9296-74ef94af5462 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.521843] env[61970]: DEBUG nova.compute.manager [req-346edeeb-9b6c-43f8-9b4c-53c0926223b4 req-a3b718e8-1eb2-44a4-8500-d1eed9b010f9 service nova] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Detach interface failed, port_id=9096db72-8476-4d1f-8fcc-cc5ddd54d61d, reason: Instance 035fb797-21c1-4af9-8e66-deee8e9d083a could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 871.527099] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355718, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099127} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.527389] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 871.528228] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d377421-c468-4b9c-a145-4113fe8b5db2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.553985] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 99200b46-7eb7-4ca2-b352-b91b4ad076af/99200b46-7eb7-4ca2-b352-b91b4ad076af.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 871.554525] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01397cc3-5ada-45fc-8840-6c0ff498d66b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.578795] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 871.578795] env[61970]: value = "task-1355720" [ 871.578795] env[61970]: _type = "Task" [ 871.578795] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.587880] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355720, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.641241] env[61970]: DEBUG nova.compute.utils [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 871.643536] env[61970]: DEBUG nova.compute.manager [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 871.643822] env[61970]: DEBUG nova.network.neutron [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 871.676852] env[61970]: DEBUG nova.network.neutron [req-49addb6f-99e7-4f39-8a62-422cec556d52 req-1d186010-fb88-47e0-bd1a-7365e2be4e4a service nova] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Updated VIF entry in instance network info cache for port 3cec35ca-1e33-4861-8a90-71573ae58d5c. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 871.677302] env[61970]: DEBUG nova.network.neutron [req-49addb6f-99e7-4f39-8a62-422cec556d52 req-1d186010-fb88-47e0-bd1a-7365e2be4e4a service nova] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Updating instance_info_cache with network_info: [{"id": "3cec35ca-1e33-4861-8a90-71573ae58d5c", "address": "fa:16:3e:5c:88:8d", "network": {"id": "a394e9eb-7879-4d19-b05f-358c85dfbe48", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-767180013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d9c85ee0fccc4373b4898f0256a69d38", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cec35ca-1e", "ovs_interfaceid": "3cec35ca-1e33-4861-8a90-71573ae58d5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.725525] env[61970]: DEBUG nova.policy [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '223ba28a2a204f8ba9e6d2e0f876ddd2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '28a078b8af9c469eb279be4da7459166', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 871.736928] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5264904e-cf5b-cd1f-02fd-4b4bf01f0770, 'name': SearchDatastore_Task, 'duration_secs': 0.00886} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.740185] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.740497] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 871.740776] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.740959] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.741157] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 871.741715] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d1264caf-d129-40f6-9624-9ee135594d38 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.751196] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 871.751403] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 871.754789] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60b6553d-246d-4a46-acca-1787a06ef501 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.761512] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 871.761512] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528b9eb9-4dc8-0ce3-3499-d7b7c3743cb5" [ 871.761512] env[61970]: _type = "Task" [ 871.761512] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.770619] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528b9eb9-4dc8-0ce3-3499-d7b7c3743cb5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.896727] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb897569-fcfa-4f45-98d7-d831ca2929a2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.902634] env[61970]: INFO nova.compute.manager [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Took 28.03 seconds to build instance. [ 871.913837] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d06236f-4957-45c1-ad19-aa4a38d9fbdd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.926051] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "refresh_cache-75762ba3-f129-424e-a7cb-962785dfab2a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.926051] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "refresh_cache-75762ba3-f129-424e-a7cb-962785dfab2a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.926051] env[61970]: DEBUG nova.network.neutron [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.926443] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355719, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.954155] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8751bf28-7f51-4cf9-9f7b-5da41fa65cd1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.962344] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a7b62cf-a2ff-4346-9c29-54ac7c939932 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.968362] env[61970]: INFO nova.compute.manager [-] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Took 1.27 seconds to deallocate network for instance. [ 871.969257] env[61970]: INFO nova.compute.manager [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] instance snapshotting [ 871.975492] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9b22963-cdc5-4066-b2f3-bf459ad9cf1f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.986696] env[61970]: DEBUG nova.compute.provider_tree [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.004471] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9547cd94-6fd3-4605-bed2-560ee80162ce {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.089479] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355720, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.150910] env[61970]: DEBUG nova.compute.manager [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 872.180935] env[61970]: DEBUG oslo_concurrency.lockutils [req-49addb6f-99e7-4f39-8a62-422cec556d52 req-1d186010-fb88-47e0-bd1a-7365e2be4e4a service nova] Releasing lock "refresh_cache-e2185ed0-2bc1-4718-b47d-57150b5e60ba" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.272728] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528b9eb9-4dc8-0ce3-3499-d7b7c3743cb5, 'name': SearchDatastore_Task, 'duration_secs': 0.010664} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.273530] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0018e00f-7ee8-42c7-9f2d-506184d5001f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.278847] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 872.278847] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52db8a62-3fdc-acaa-a4ff-147492ac1b30" [ 872.278847] env[61970]: _type = "Task" [ 872.278847] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.286401] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52db8a62-3fdc-acaa-a4ff-147492ac1b30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.301382] env[61970]: DEBUG nova.network.neutron [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Successfully created port: 2545e2b0-a432-44ae-a287-0b68ceab7031 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 872.413445] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3baeecc2-bb8d-4887-98e3-54a29580be3f tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.930s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.413708] env[61970]: DEBUG oslo_vmware.api [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355719, 'name': PowerOnVM_Task, 'duration_secs': 0.874907} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.413947] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 872.414178] env[61970]: DEBUG nova.compute.manager [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 872.414891] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1919b86-dadf-4868-9e9c-71a482f5d7fd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.458678] env[61970]: DEBUG nova.network.neutron [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 872.479111] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.490161] env[61970]: DEBUG nova.scheduler.client.report [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 872.514748] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Creating Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 872.515391] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d7c5bdd9-860e-4c04-98cb-e97a18b4420b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.523960] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for the task: (returnval){ [ 872.523960] env[61970]: value = "task-1355721" [ 872.523960] env[61970]: _type = "Task" [ 872.523960] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.536150] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355721, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.591096] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355720, 'name': ReconfigVM_Task, 'duration_secs': 0.677874} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.591407] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 99200b46-7eb7-4ca2-b352-b91b4ad076af/99200b46-7eb7-4ca2-b352-b91b4ad076af.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 872.592079] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd24a77d-0ee8-4f19-bb37-8e30f3655ddf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.595017] env[61970]: DEBUG nova.network.neutron [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Updating instance_info_cache with network_info: [{"id": "32711b6a-a9cc-472d-a634-e3b60711bffe", "address": "fa:16:3e:68:3d:bc", "network": {"id": "cba185ff-bfa3-4547-b7b0-638c3e79a24c", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-864852491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5c64cb9232fb413cbd7627dcf077e9ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32711b6a-a9", "ovs_interfaceid": "32711b6a-a9cc-472d-a634-e3b60711bffe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.598858] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 872.598858] env[61970]: value = "task-1355722" [ 872.598858] env[61970]: _type = "Task" [ 872.598858] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.606272] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355722, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.775018] env[61970]: DEBUG nova.compute.manager [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Received event network-vif-plugged-32711b6a-a9cc-472d-a634-e3b60711bffe {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.775329] env[61970]: DEBUG oslo_concurrency.lockutils [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] Acquiring lock "75762ba3-f129-424e-a7cb-962785dfab2a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.775584] env[61970]: DEBUG oslo_concurrency.lockutils [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] Lock "75762ba3-f129-424e-a7cb-962785dfab2a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.775786] env[61970]: DEBUG oslo_concurrency.lockutils [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] Lock "75762ba3-f129-424e-a7cb-962785dfab2a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.775999] env[61970]: DEBUG nova.compute.manager [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] No waiting events found dispatching network-vif-plugged-32711b6a-a9cc-472d-a634-e3b60711bffe {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 872.776226] env[61970]: WARNING nova.compute.manager [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Received unexpected event network-vif-plugged-32711b6a-a9cc-472d-a634-e3b60711bffe for instance with vm_state building and task_state spawning. [ 872.776560] env[61970]: DEBUG nova.compute.manager [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Received event network-changed-32711b6a-a9cc-472d-a634-e3b60711bffe {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.776796] env[61970]: DEBUG nova.compute.manager [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Refreshing instance network info cache due to event network-changed-32711b6a-a9cc-472d-a634-e3b60711bffe. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 872.776965] env[61970]: DEBUG oslo_concurrency.lockutils [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] Acquiring lock "refresh_cache-75762ba3-f129-424e-a7cb-962785dfab2a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.791981] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52db8a62-3fdc-acaa-a4ff-147492ac1b30, 'name': SearchDatastore_Task, 'duration_secs': 0.024352} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.792473] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.792733] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] e2185ed0-2bc1-4718-b47d-57150b5e60ba/e2185ed0-2bc1-4718-b47d-57150b5e60ba.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 872.793306] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1cbb7d12-fcc4-43cb-a680-eb8a41de03e3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.802021] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 872.802021] env[61970]: value = "task-1355723" [ 872.802021] env[61970]: _type = "Task" [ 872.802021] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.809349] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355723, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.930398] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.995475] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.858s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.998943] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.383s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.998943] env[61970]: DEBUG nova.objects.instance [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lazy-loading 'resources' on Instance uuid e2f51299-56fe-46cc-9e08-2246178b9db7 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.021495] env[61970]: INFO nova.scheduler.client.report [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Deleted allocations for instance 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0 [ 873.034796] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355721, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.102604] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "refresh_cache-75762ba3-f129-424e-a7cb-962785dfab2a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.103106] env[61970]: DEBUG nova.compute.manager [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Instance network_info: |[{"id": "32711b6a-a9cc-472d-a634-e3b60711bffe", "address": "fa:16:3e:68:3d:bc", "network": {"id": "cba185ff-bfa3-4547-b7b0-638c3e79a24c", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-864852491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5c64cb9232fb413cbd7627dcf077e9ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32711b6a-a9", "ovs_interfaceid": "32711b6a-a9cc-472d-a634-e3b60711bffe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 873.103351] env[61970]: DEBUG oslo_concurrency.lockutils [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] Acquired lock "refresh_cache-75762ba3-f129-424e-a7cb-962785dfab2a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.103538] env[61970]: DEBUG nova.network.neutron [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Refreshing network info cache for port 32711b6a-a9cc-472d-a634-e3b60711bffe {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 873.104699] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:3d:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '40c947c4-f471-4d48-8e43-fee54198107e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '32711b6a-a9cc-472d-a634-e3b60711bffe', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 873.112714] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Creating folder: Project (5c64cb9232fb413cbd7627dcf077e9ef). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 873.116273] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2aa0ee86-6c2a-4c96-812b-23fb7b7713a5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.121555] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355722, 'name': Rename_Task, 'duration_secs': 0.290019} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.122132] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 873.122371] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9cdaf914-0f6c-48c1-a2a5-28e5442cf578 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.127673] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 873.127673] env[61970]: value = "task-1355725" [ 873.127673] env[61970]: _type = "Task" [ 873.127673] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.131627] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Created folder: Project (5c64cb9232fb413cbd7627dcf077e9ef) in parent group-v288740. [ 873.132032] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Creating folder: Instances. Parent ref: group-v288808. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 873.132364] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9aaa044b-ba6a-4c7f-8ad0-3dac97309dc3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.136935] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355725, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.143810] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Created folder: Instances in parent group-v288808. [ 873.144040] env[61970]: DEBUG oslo.service.loopingcall [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.144218] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 873.144425] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed8bd8f8-8f09-4901-925c-0994d14016e4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.162455] env[61970]: DEBUG nova.compute.manager [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 873.169581] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 873.169581] env[61970]: value = "task-1355727" [ 873.169581] env[61970]: _type = "Task" [ 873.169581] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.178143] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355727, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.191033] env[61970]: DEBUG nova.virt.hardware [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 873.191033] env[61970]: DEBUG nova.virt.hardware [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 873.191253] env[61970]: DEBUG nova.virt.hardware [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 873.192208] env[61970]: DEBUG nova.virt.hardware [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 873.192208] env[61970]: DEBUG nova.virt.hardware [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 873.192208] env[61970]: DEBUG nova.virt.hardware [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 873.192208] env[61970]: DEBUG nova.virt.hardware [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 873.192208] env[61970]: DEBUG nova.virt.hardware [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 873.192437] env[61970]: DEBUG nova.virt.hardware [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 873.192437] env[61970]: DEBUG nova.virt.hardware [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 873.192515] env[61970]: DEBUG nova.virt.hardware [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 873.193537] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e69131-692b-4378-9917-33fbbb2e7f7d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.203828] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fbee8a4-3bc0-4124-a99e-a159dd68630d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.310836] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355723, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.403305] env[61970]: DEBUG nova.network.neutron [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Updated VIF entry in instance network info cache for port 32711b6a-a9cc-472d-a634-e3b60711bffe. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 873.403708] env[61970]: DEBUG nova.network.neutron [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Updating instance_info_cache with network_info: [{"id": "32711b6a-a9cc-472d-a634-e3b60711bffe", "address": "fa:16:3e:68:3d:bc", "network": {"id": "cba185ff-bfa3-4547-b7b0-638c3e79a24c", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-864852491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5c64cb9232fb413cbd7627dcf077e9ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32711b6a-a9", "ovs_interfaceid": "32711b6a-a9cc-472d-a634-e3b60711bffe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.534459] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0a6712cc-7b4e-4b13-adb0-ee656abdb1a7 tempest-ServerShowV247Test-1223937591 tempest-ServerShowV247Test-1223937591-project-member] Lock "7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.167s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.539478] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355721, 'name': CreateSnapshot_Task, 'duration_secs': 0.842336} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.540041] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Created Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 873.540866] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c153142-be09-48e1-9a50-a86aab2aad1c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.641685] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355725, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.683346] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355727, 'name': CreateVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.747237] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2861a8f9-e2b1-417e-b625-cf975ec9ddcd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.755166] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290a9525-b953-47cc-9b93-871d54d5b3d5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.787629] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7b7ec4-005a-4b2e-9130-956b5378345c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.795421] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd764c53-13f2-4a00-9b43-c2f3b1d02ccf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.809396] env[61970]: DEBUG nova.compute.provider_tree [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.819351] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355723, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.907157] env[61970]: DEBUG oslo_concurrency.lockutils [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] Releasing lock "refresh_cache-75762ba3-f129-424e-a7cb-962785dfab2a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.907757] env[61970]: DEBUG nova.compute.manager [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Received event network-changed-7bfd07fd-97bc-42b5-9571-0807d864f9bc {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.907839] env[61970]: DEBUG nova.compute.manager [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Refreshing instance network info cache due to event network-changed-7bfd07fd-97bc-42b5-9571-0807d864f9bc. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 873.908119] env[61970]: DEBUG oslo_concurrency.lockutils [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] Acquiring lock "refresh_cache-15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.908970] env[61970]: DEBUG oslo_concurrency.lockutils [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] Acquired lock "refresh_cache-15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.908970] env[61970]: DEBUG nova.network.neutron [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Refreshing network info cache for port 7bfd07fd-97bc-42b5-9571-0807d864f9bc {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 874.014695] env[61970]: DEBUG nova.network.neutron [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Successfully updated port: 2545e2b0-a432-44ae-a287-0b68ceab7031 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 874.068479] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Creating linked-clone VM from snapshot {{(pid=61970) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 874.070733] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d7819450-35f8-4ed6-9c3c-bdfba1d7bcf5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.078827] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for the task: (returnval){ [ 874.078827] env[61970]: value = "task-1355728" [ 874.078827] env[61970]: _type = "Task" [ 874.078827] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.087012] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355728, 'name': CloneVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.087872] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Acquiring lock "e1415314-f137-4ee3-a065-3d875ef9e7ff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.087872] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Lock "e1415314-f137-4ee3-a065-3d875ef9e7ff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.087872] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Acquiring lock "e1415314-f137-4ee3-a065-3d875ef9e7ff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.088321] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Lock "e1415314-f137-4ee3-a065-3d875ef9e7ff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.088321] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Lock "e1415314-f137-4ee3-a065-3d875ef9e7ff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.091327] env[61970]: INFO nova.compute.manager [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Terminating instance [ 874.091878] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Acquiring lock "refresh_cache-e1415314-f137-4ee3-a065-3d875ef9e7ff" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.092179] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Acquired lock "refresh_cache-e1415314-f137-4ee3-a065-3d875ef9e7ff" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.092470] env[61970]: DEBUG nova.network.neutron [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 874.139855] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355725, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.181146] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355727, 'name': CreateVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.316282] env[61970]: DEBUG nova.scheduler.client.report [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.324326] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355723, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.519634] env[61970]: DEBUG oslo_concurrency.lockutils [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "refresh_cache-0304640a-6c46-4f57-a8b8-5058da2e3de4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.519634] env[61970]: DEBUG oslo_concurrency.lockutils [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "refresh_cache-0304640a-6c46-4f57-a8b8-5058da2e3de4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.519634] env[61970]: DEBUG nova.network.neutron [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 874.589915] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355728, 'name': CloneVM_Task} progress is 94%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.612892] env[61970]: DEBUG nova.network.neutron [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 874.646234] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355725, 'name': PowerOnVM_Task, 'duration_secs': 1.188315} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.646234] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 874.646234] env[61970]: INFO nova.compute.manager [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Took 8.65 seconds to spawn the instance on the hypervisor. [ 874.646234] env[61970]: DEBUG nova.compute.manager [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 874.647945] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb0b1fb5-5729-4066-b7d4-ae0bc1058cd6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.685917] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355727, 'name': CreateVM_Task} progress is 25%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.715056] env[61970]: DEBUG nova.network.neutron [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.767852] env[61970]: DEBUG nova.network.neutron [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Updated VIF entry in instance network info cache for port 7bfd07fd-97bc-42b5-9571-0807d864f9bc. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 874.768447] env[61970]: DEBUG nova.network.neutron [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Updating instance_info_cache with network_info: [{"id": "7bfd07fd-97bc-42b5-9571-0807d864f9bc", "address": "fa:16:3e:89:e6:54", "network": {"id": "42def9ae-4eed-41e5-a538-81eaced2cdd4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2047751481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19dae2316adb4dc980bd972c8c40c034", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c1b8689-a9b4-4972-beb9-6a1c8de1dc88", "external-id": "nsx-vlan-transportzone-455", "segmentation_id": 455, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bfd07fd-97", "ovs_interfaceid": "7bfd07fd-97bc-42b5-9571-0807d864f9bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.800117] env[61970]: DEBUG nova.compute.manager [req-95fde27d-3e0c-4d0c-b3dc-d54a7ca18901 req-9091b12d-8824-4a8c-9898-c67785a872df service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Received event network-vif-plugged-2545e2b0-a432-44ae-a287-0b68ceab7031 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.800419] env[61970]: DEBUG oslo_concurrency.lockutils [req-95fde27d-3e0c-4d0c-b3dc-d54a7ca18901 req-9091b12d-8824-4a8c-9898-c67785a872df service nova] Acquiring lock "0304640a-6c46-4f57-a8b8-5058da2e3de4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.800644] env[61970]: DEBUG oslo_concurrency.lockutils [req-95fde27d-3e0c-4d0c-b3dc-d54a7ca18901 req-9091b12d-8824-4a8c-9898-c67785a872df service nova] Lock "0304640a-6c46-4f57-a8b8-5058da2e3de4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.800836] env[61970]: DEBUG oslo_concurrency.lockutils [req-95fde27d-3e0c-4d0c-b3dc-d54a7ca18901 req-9091b12d-8824-4a8c-9898-c67785a872df service nova] Lock "0304640a-6c46-4f57-a8b8-5058da2e3de4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.801196] env[61970]: DEBUG nova.compute.manager [req-95fde27d-3e0c-4d0c-b3dc-d54a7ca18901 req-9091b12d-8824-4a8c-9898-c67785a872df service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] No waiting events found dispatching network-vif-plugged-2545e2b0-a432-44ae-a287-0b68ceab7031 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 874.801460] env[61970]: WARNING nova.compute.manager [req-95fde27d-3e0c-4d0c-b3dc-d54a7ca18901 req-9091b12d-8824-4a8c-9898-c67785a872df service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Received unexpected event network-vif-plugged-2545e2b0-a432-44ae-a287-0b68ceab7031 for instance with vm_state building and task_state spawning. [ 874.801677] env[61970]: DEBUG nova.compute.manager [req-95fde27d-3e0c-4d0c-b3dc-d54a7ca18901 req-9091b12d-8824-4a8c-9898-c67785a872df service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Received event network-changed-2545e2b0-a432-44ae-a287-0b68ceab7031 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.801921] env[61970]: DEBUG nova.compute.manager [req-95fde27d-3e0c-4d0c-b3dc-d54a7ca18901 req-9091b12d-8824-4a8c-9898-c67785a872df service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Refreshing instance network info cache due to event network-changed-2545e2b0-a432-44ae-a287-0b68ceab7031. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 874.802177] env[61970]: DEBUG oslo_concurrency.lockutils [req-95fde27d-3e0c-4d0c-b3dc-d54a7ca18901 req-9091b12d-8824-4a8c-9898-c67785a872df service nova] Acquiring lock "refresh_cache-0304640a-6c46-4f57-a8b8-5058da2e3de4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.821899] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355723, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.882522} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.822179] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] e2185ed0-2bc1-4718-b47d-57150b5e60ba/e2185ed0-2bc1-4718-b47d-57150b5e60ba.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 874.822467] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 874.822732] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-531175a4-8010-4f03-bd1e-c34cb57a0799 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.825542] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.828072] env[61970]: DEBUG oslo_concurrency.lockutils [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.379s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.828315] env[61970]: DEBUG nova.objects.instance [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lazy-loading 'resources' on Instance uuid d09dbde2-5daa-4ae9-9fd5-bfbb03fba136 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.834356] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 874.834356] env[61970]: value = "task-1355729" [ 874.834356] env[61970]: _type = "Task" [ 874.834356] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.844581] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355729, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.856325] env[61970]: INFO nova.scheduler.client.report [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Deleted allocations for instance e2f51299-56fe-46cc-9e08-2246178b9db7 [ 875.058439] env[61970]: DEBUG nova.network.neutron [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 875.089131] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355728, 'name': CloneVM_Task} progress is 94%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.173084] env[61970]: INFO nova.compute.manager [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Took 23.12 seconds to build instance. [ 875.186387] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355727, 'name': CreateVM_Task, 'duration_secs': 1.986245} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.186723] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 875.187716] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.188124] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.188481] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 875.189059] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14761708-d659-416e-b01e-2490c980adae {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.194716] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 875.194716] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ffe63b-0920-e828-80f5-4e643275a6aa" [ 875.194716] env[61970]: _type = "Task" [ 875.194716] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.205221] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ffe63b-0920-e828-80f5-4e643275a6aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.217626] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Releasing lock "refresh_cache-e1415314-f137-4ee3-a065-3d875ef9e7ff" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.218577] env[61970]: DEBUG nova.compute.manager [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 875.218577] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 875.219859] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8032bbd7-5d2c-4258-92aa-1cf54054d04b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.228069] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 875.229430] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95969496-77b5-4cdf-bd70-5629048f27f7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.234499] env[61970]: DEBUG nova.network.neutron [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Updating instance_info_cache with network_info: [{"id": "2545e2b0-a432-44ae-a287-0b68ceab7031", "address": "fa:16:3e:a6:77:ea", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2545e2b0-a4", "ovs_interfaceid": "2545e2b0-a432-44ae-a287-0b68ceab7031", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.241413] env[61970]: DEBUG oslo_vmware.api [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 875.241413] env[61970]: value = "task-1355730" [ 875.241413] env[61970]: _type = "Task" [ 875.241413] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.251181] env[61970]: DEBUG oslo_vmware.api [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355730, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.270599] env[61970]: DEBUG oslo_concurrency.lockutils [req-6f3df547-d7de-4201-a842-ab5e50b8b739 req-14fe8608-00ba-4b0f-bbe3-bb13688d153e service nova] Releasing lock "refresh_cache-15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.344606] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355729, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.131853} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.344951] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 875.345878] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-207c3026-bdd3-43d8-9abe-5b293803e1fe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.369895] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] e2185ed0-2bc1-4718-b47d-57150b5e60ba/e2185ed0-2bc1-4718-b47d-57150b5e60ba.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 875.374962] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c4ec1291-a614-4414-87d6-a13239291a0f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.391168] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dc02252e-cd1a-4e03-b069-31ae948a7197 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "e2f51299-56fe-46cc-9e08-2246178b9db7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.913s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.398471] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 875.398471] env[61970]: value = "task-1355731" [ 875.398471] env[61970]: _type = "Task" [ 875.398471] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.411695] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355731, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.589662] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355728, 'name': CloneVM_Task} progress is 94%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.608274] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae1940d-988e-47cc-a6c2-44ef99440670 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.617358] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3e6052b-34bd-4592-b2c7-679981a95eb5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.650530] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8fdb366-039f-43f9-9873-ef3df15d813c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.660330] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47b92370-dc1f-4651-8b9c-bcf150ca8541 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.678759] env[61970]: DEBUG nova.compute.provider_tree [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.680316] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "99200b46-7eb7-4ca2-b352-b91b4ad076af" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.585s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.706085] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ffe63b-0920-e828-80f5-4e643275a6aa, 'name': SearchDatastore_Task, 'duration_secs': 0.014017} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.706356] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.706589] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 875.707027] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.707027] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.707141] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 875.707400] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d934dbf8-b8c6-4bc4-9491-ee858ce6a82a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.716082] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 875.716270] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 875.716991] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38f19fe5-2b55-4f28-a927-25cc93ea94c7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.722533] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 875.722533] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f49854-e7ee-72c0-5b80-395d6301fc4a" [ 875.722533] env[61970]: _type = "Task" [ 875.722533] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.730762] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f49854-e7ee-72c0-5b80-395d6301fc4a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.737342] env[61970]: DEBUG oslo_concurrency.lockutils [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "refresh_cache-0304640a-6c46-4f57-a8b8-5058da2e3de4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.737639] env[61970]: DEBUG nova.compute.manager [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Instance network_info: |[{"id": "2545e2b0-a432-44ae-a287-0b68ceab7031", "address": "fa:16:3e:a6:77:ea", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2545e2b0-a4", "ovs_interfaceid": "2545e2b0-a432-44ae-a287-0b68ceab7031", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 875.737967] env[61970]: DEBUG oslo_concurrency.lockutils [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "6345d239-fbae-4f3a-9f61-79e10539ff0b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.738190] env[61970]: DEBUG oslo_concurrency.lockutils [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "6345d239-fbae-4f3a-9f61-79e10539ff0b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.738419] env[61970]: DEBUG oslo_concurrency.lockutils [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "6345d239-fbae-4f3a-9f61-79e10539ff0b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.738566] env[61970]: DEBUG oslo_concurrency.lockutils [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "6345d239-fbae-4f3a-9f61-79e10539ff0b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.738726] env[61970]: DEBUG oslo_concurrency.lockutils [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "6345d239-fbae-4f3a-9f61-79e10539ff0b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.740355] env[61970]: DEBUG oslo_concurrency.lockutils [req-95fde27d-3e0c-4d0c-b3dc-d54a7ca18901 req-9091b12d-8824-4a8c-9898-c67785a872df service nova] Acquired lock "refresh_cache-0304640a-6c46-4f57-a8b8-5058da2e3de4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.740455] env[61970]: DEBUG nova.network.neutron [req-95fde27d-3e0c-4d0c-b3dc-d54a7ca18901 req-9091b12d-8824-4a8c-9898-c67785a872df service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Refreshing network info cache for port 2545e2b0-a432-44ae-a287-0b68ceab7031 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 875.741703] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a6:77:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b356db78-99c7-4464-822c-fc7e193f7878', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2545e2b0-a432-44ae-a287-0b68ceab7031', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.749092] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Creating folder: Project (28a078b8af9c469eb279be4da7459166). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 875.749734] env[61970]: INFO nova.compute.manager [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Terminating instance [ 875.751623] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ab9a5be5-85bb-470a-b083-106e5f69660c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.753812] env[61970]: DEBUG nova.compute.manager [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 875.754011] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 875.758427] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92459096-7478-4bc5-a003-2bf3bcb15c12 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.767263] env[61970]: DEBUG oslo_vmware.api [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355730, 'name': PowerOffVM_Task, 'duration_secs': 0.241149} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.769696] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 875.769871] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 875.770147] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 875.771213] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-59186d1c-90ed-4afc-82cb-7814fdca990b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.772589] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4182c78d-e121-4cad-9c64-66a684decebe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.774162] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Created folder: Project (28a078b8af9c469eb279be4da7459166) in parent group-v288740. [ 875.774344] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Creating folder: Instances. Parent ref: group-v288813. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 875.774563] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f7961e7d-b2fe-4f83-839a-93285f8bc38d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.781624] env[61970]: DEBUG oslo_vmware.api [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 875.781624] env[61970]: value = "task-1355733" [ 875.781624] env[61970]: _type = "Task" [ 875.781624] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.786795] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Created folder: Instances in parent group-v288813. [ 875.787027] env[61970]: DEBUG oslo.service.loopingcall [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.790471] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 875.790755] env[61970]: DEBUG oslo_vmware.api [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355733, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.790966] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-34cf4489-2297-4704-ac41-2388e887e34d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.806984] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 875.807216] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 875.807402] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Deleting the datastore file [datastore2] e1415314-f137-4ee3-a065-3d875ef9e7ff {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 875.808065] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f66e3f0b-5c8d-4f85-b0f7-91844826ef71 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.811938] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 875.811938] env[61970]: value = "task-1355736" [ 875.811938] env[61970]: _type = "Task" [ 875.811938] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.816944] env[61970]: DEBUG oslo_vmware.api [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for the task: (returnval){ [ 875.816944] env[61970]: value = "task-1355737" [ 875.816944] env[61970]: _type = "Task" [ 875.816944] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.822582] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355736, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.827387] env[61970]: DEBUG oslo_vmware.api [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355737, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.909767] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355731, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.091951] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355728, 'name': CloneVM_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.182491] env[61970]: DEBUG nova.scheduler.client.report [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.233571] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f49854-e7ee-72c0-5b80-395d6301fc4a, 'name': SearchDatastore_Task, 'duration_secs': 0.009712} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.234788] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b7c3539-f72f-4b0d-aeb3-6c29b4121e85 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.239692] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 876.239692] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52cfab43-46d4-97cc-dae8-aaef83504d7a" [ 876.239692] env[61970]: _type = "Task" [ 876.239692] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.246873] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52cfab43-46d4-97cc-dae8-aaef83504d7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.295132] env[61970]: DEBUG oslo_vmware.api [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355733, 'name': PowerOffVM_Task, 'duration_secs': 0.27287} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.295132] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 876.295132] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 876.295132] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3ac2a020-8464-4607-a0c6-17744078e0a6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.326234] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355736, 'name': CreateVM_Task, 'duration_secs': 0.356687} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.329068] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 876.329759] env[61970]: DEBUG oslo_concurrency.lockutils [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.329924] env[61970]: DEBUG oslo_concurrency.lockutils [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.330261] env[61970]: DEBUG oslo_concurrency.lockutils [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 876.333973] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1a43316-7487-4b07-a76f-d19d1e98a0bb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.335645] env[61970]: DEBUG oslo_vmware.api [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Task: {'id': task-1355737, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175312} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.335885] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 876.336076] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 876.336263] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 876.336428] env[61970]: INFO nova.compute.manager [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Took 1.12 seconds to destroy the instance on the hypervisor. [ 876.336680] env[61970]: DEBUG oslo.service.loopingcall [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.337249] env[61970]: DEBUG nova.compute.manager [-] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 876.337371] env[61970]: DEBUG nova.network.neutron [-] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 876.340396] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 876.340396] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52173ef4-fcc8-541f-52fb-53b82e82d67f" [ 876.340396] env[61970]: _type = "Task" [ 876.340396] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.348723] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52173ef4-fcc8-541f-52fb-53b82e82d67f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.355667] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 876.355884] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 876.357430] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Deleting the datastore file [datastore2] 6345d239-fbae-4f3a-9f61-79e10539ff0b {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 876.357430] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5e0b1a88-6ac0-4f7f-9ed0-1cc09cb68b14 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.363874] env[61970]: DEBUG oslo_vmware.api [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 876.363874] env[61970]: value = "task-1355739" [ 876.363874] env[61970]: _type = "Task" [ 876.363874] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.372062] env[61970]: DEBUG oslo_vmware.api [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355739, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.372764] env[61970]: DEBUG nova.network.neutron [-] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 876.409888] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355731, 'name': ReconfigVM_Task, 'duration_secs': 0.643268} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.410222] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Reconfigured VM instance instance-0000004a to attach disk [datastore1] e2185ed0-2bc1-4718-b47d-57150b5e60ba/e2185ed0-2bc1-4718-b47d-57150b5e60ba.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.410901] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-78b9082f-8f94-4ab9-a7a0-06949ed96291 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.417081] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 876.417081] env[61970]: value = "task-1355740" [ 876.417081] env[61970]: _type = "Task" [ 876.417081] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.427398] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355740, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.490977] env[61970]: DEBUG nova.network.neutron [req-95fde27d-3e0c-4d0c-b3dc-d54a7ca18901 req-9091b12d-8824-4a8c-9898-c67785a872df service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Updated VIF entry in instance network info cache for port 2545e2b0-a432-44ae-a287-0b68ceab7031. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 876.491449] env[61970]: DEBUG nova.network.neutron [req-95fde27d-3e0c-4d0c-b3dc-d54a7ca18901 req-9091b12d-8824-4a8c-9898-c67785a872df service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Updating instance_info_cache with network_info: [{"id": "2545e2b0-a432-44ae-a287-0b68ceab7031", "address": "fa:16:3e:a6:77:ea", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2545e2b0-a4", "ovs_interfaceid": "2545e2b0-a432-44ae-a287-0b68ceab7031", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.593751] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355728, 'name': CloneVM_Task, 'duration_secs': 2.094507} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.593751] env[61970]: INFO nova.virt.vmwareapi.vmops [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Created linked-clone VM from snapshot [ 876.593751] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eda2869-59c0-43f6-be18-3b9a949e3ad3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.600007] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Uploading image 34020495-c8a7-4322-81b7-a1494c1485f7 {{(pid=61970) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 876.619975] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Destroying the VM {{(pid=61970) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 876.620319] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-58b8c8ca-beaf-4593-96b0-edca96bbb89f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.626697] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for the task: (returnval){ [ 876.626697] env[61970]: value = "task-1355741" [ 876.626697] env[61970]: _type = "Task" [ 876.626697] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.634763] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355741, 'name': Destroy_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.688255] env[61970]: DEBUG oslo_concurrency.lockutils [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.860s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.690624] env[61970]: DEBUG oslo_concurrency.lockutils [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.458s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.690859] env[61970]: DEBUG nova.objects.instance [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lazy-loading 'resources' on Instance uuid b19cb516-b163-4bed-ba5b-139a0a18fc05 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 876.714422] env[61970]: INFO nova.scheduler.client.report [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleted allocations for instance d09dbde2-5daa-4ae9-9fd5-bfbb03fba136 [ 876.750642] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52cfab43-46d4-97cc-dae8-aaef83504d7a, 'name': SearchDatastore_Task, 'duration_secs': 0.009591} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.750905] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.751186] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] 75762ba3-f129-424e-a7cb-962785dfab2a/75762ba3-f129-424e-a7cb-962785dfab2a.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 876.751744] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5bc125a1-5424-4c3a-a12e-8376bc9c505b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.759688] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 876.759688] env[61970]: value = "task-1355742" [ 876.759688] env[61970]: _type = "Task" [ 876.759688] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.767205] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355742, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.852845] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52173ef4-fcc8-541f-52fb-53b82e82d67f, 'name': SearchDatastore_Task, 'duration_secs': 0.015816} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.853030] env[61970]: DEBUG oslo_concurrency.lockutils [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.853327] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.853503] env[61970]: DEBUG oslo_concurrency.lockutils [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.853665] env[61970]: DEBUG oslo_concurrency.lockutils [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.853845] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.854114] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca182ba6-c99c-4523-89e6-32bbebfca2ee {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.862191] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.862405] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 876.863239] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b136a974-68af-4827-b67c-f4b2eb5430d2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.874213] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 876.874213] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52052487-7b51-f0cd-9183-78720aa430ce" [ 876.874213] env[61970]: _type = "Task" [ 876.874213] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.876809] env[61970]: DEBUG nova.network.neutron [-] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.878116] env[61970]: DEBUG oslo_vmware.api [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355739, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177474} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.881408] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 876.881633] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 876.881815] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 876.882056] env[61970]: INFO nova.compute.manager [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Took 1.13 seconds to destroy the instance on the hypervisor. [ 876.882224] env[61970]: DEBUG oslo.service.loopingcall [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.882418] env[61970]: DEBUG nova.compute.manager [-] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 876.882520] env[61970]: DEBUG nova.network.neutron [-] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 876.889975] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52052487-7b51-f0cd-9183-78720aa430ce, 'name': SearchDatastore_Task, 'duration_secs': 0.010364} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.890935] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75f1525d-18cc-403e-a13e-bd20de827362 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.896517] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 876.896517] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5265a38a-ad30-b325-c368-fc5fbc438e90" [ 876.896517] env[61970]: _type = "Task" [ 876.896517] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.904514] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5265a38a-ad30-b325-c368-fc5fbc438e90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.927234] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355740, 'name': Rename_Task, 'duration_secs': 0.185357} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.927519] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 876.927773] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff605bb8-205b-4abe-846e-fc9d380f71f5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.933956] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 876.933956] env[61970]: value = "task-1355743" [ 876.933956] env[61970]: _type = "Task" [ 876.933956] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.942737] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355743, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.994879] env[61970]: DEBUG oslo_concurrency.lockutils [req-95fde27d-3e0c-4d0c-b3dc-d54a7ca18901 req-9091b12d-8824-4a8c-9898-c67785a872df service nova] Releasing lock "refresh_cache-0304640a-6c46-4f57-a8b8-5058da2e3de4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.142399] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355741, 'name': Destroy_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.226722] env[61970]: DEBUG oslo_concurrency.lockutils [None req-497f4f91-b012-448c-be45-697156838c0c tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "d09dbde2-5daa-4ae9-9fd5-bfbb03fba136" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.199s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.273114] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355742, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.277181] env[61970]: DEBUG nova.compute.manager [req-ffb7e690-3637-445b-afc1-849f25d55dab req-87e0f975-e504-4484-9055-d861d73a3e84 service nova] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Received event network-vif-deleted-c2860647-20d9-411e-9d79-e26336519408 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.278253] env[61970]: INFO nova.compute.manager [req-ffb7e690-3637-445b-afc1-849f25d55dab req-87e0f975-e504-4484-9055-d861d73a3e84 service nova] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Neutron deleted interface c2860647-20d9-411e-9d79-e26336519408; detaching it from the instance and deleting it from the info cache [ 877.278520] env[61970]: DEBUG nova.network.neutron [req-ffb7e690-3637-445b-afc1-849f25d55dab req-87e0f975-e504-4484-9055-d861d73a3e84 service nova] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.380033] env[61970]: INFO nova.compute.manager [-] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Took 1.04 seconds to deallocate network for instance. [ 877.406749] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5265a38a-ad30-b325-c368-fc5fbc438e90, 'name': SearchDatastore_Task, 'duration_secs': 0.023195} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.407017] env[61970]: DEBUG oslo_concurrency.lockutils [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.407278] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] 0304640a-6c46-4f57-a8b8-5058da2e3de4/0304640a-6c46-4f57-a8b8-5058da2e3de4.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 877.407545] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-220cd0c5-4fa6-4be1-8175-7cea442531e2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.414698] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 877.414698] env[61970]: value = "task-1355744" [ 877.414698] env[61970]: _type = "Task" [ 877.414698] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.424433] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355744, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.440748] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd75062-7393-4740-92bd-8f5c13e0e7a8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.446601] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355743, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.451344] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f4ce90-4adf-4ce9-b359-99e327f44827 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.484315] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e1af4eb-0a87-4471-a619-4f91630ec147 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.492463] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a77b2f8f-7d6a-44a3-ba3b-723a05f5f1ab {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.507076] env[61970]: DEBUG nova.compute.provider_tree [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 877.642801] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355741, 'name': Destroy_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.738141] env[61970]: DEBUG nova.network.neutron [-] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.773047] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355742, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.582029} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.773237] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] 75762ba3-f129-424e-a7cb-962785dfab2a/75762ba3-f129-424e-a7cb-962785dfab2a.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 877.773472] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 877.773745] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b6c07149-8058-413e-82f2-cfd5f2a05e72 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.780741] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 877.780741] env[61970]: value = "task-1355745" [ 877.780741] env[61970]: _type = "Task" [ 877.780741] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.785239] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b11c57b6-7eaa-433b-83c7-cb53bae7fca5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.792514] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355745, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.796937] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e204c17-838f-45d4-b1f0-2a2e0ab11d5b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.834683] env[61970]: DEBUG nova.compute.manager [req-ffb7e690-3637-445b-afc1-849f25d55dab req-87e0f975-e504-4484-9055-d861d73a3e84 service nova] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Detach interface failed, port_id=c2860647-20d9-411e-9d79-e26336519408, reason: Instance 6345d239-fbae-4f3a-9f61-79e10539ff0b could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 877.888141] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.925140] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355744, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.944551] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355743, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.011546] env[61970]: DEBUG nova.scheduler.client.report [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 878.140508] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355741, 'name': Destroy_Task} progress is 33%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.241609] env[61970]: INFO nova.compute.manager [-] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Took 1.36 seconds to deallocate network for instance. [ 878.292620] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355745, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.116898} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.293058] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 878.294876] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682f4f41-e67d-48a8-aca4-974fae046dfe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.321216] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 75762ba3-f129-424e-a7cb-962785dfab2a/75762ba3-f129-424e-a7cb-962785dfab2a.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 878.322499] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ceaa1de-52bd-4015-ba4b-b046e955b7cf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.344956] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 878.344956] env[61970]: value = "task-1355746" [ 878.344956] env[61970]: _type = "Task" [ 878.344956] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.353019] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355746, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.425621] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355744, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.52385} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.425872] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] 0304640a-6c46-4f57-a8b8-5058da2e3de4/0304640a-6c46-4f57-a8b8-5058da2e3de4.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 878.426095] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 878.426362] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e778034b-edb6-46de-8afd-30f543f62e9f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.433263] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 878.433263] env[61970]: value = "task-1355747" [ 878.433263] env[61970]: _type = "Task" [ 878.433263] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.443556] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355747, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.446651] env[61970]: DEBUG oslo_vmware.api [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355743, 'name': PowerOnVM_Task, 'duration_secs': 1.335818} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.446868] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 878.447080] env[61970]: INFO nova.compute.manager [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Took 10.06 seconds to spawn the instance on the hypervisor. [ 878.447261] env[61970]: DEBUG nova.compute.manager [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.448046] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9524526-eaf2-46e7-b95d-ec0b1e5e0ead {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.516584] env[61970]: DEBUG oslo_concurrency.lockutils [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.826s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.519322] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 7.313s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.544667] env[61970]: INFO nova.scheduler.client.report [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Deleted allocations for instance b19cb516-b163-4bed-ba5b-139a0a18fc05 [ 878.643564] env[61970]: DEBUG oslo_concurrency.lockutils [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.643865] env[61970]: DEBUG oslo_concurrency.lockutils [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.652550] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355741, 'name': Destroy_Task, 'duration_secs': 1.561694} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.653023] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Destroyed the VM [ 878.653392] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Deleting Snapshot of the VM instance {{(pid=61970) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 878.653646] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-20d9528f-dd51-4abf-b56c-157cef27d787 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.663955] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for the task: (returnval){ [ 878.663955] env[61970]: value = "task-1355748" [ 878.663955] env[61970]: _type = "Task" [ 878.663955] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.673991] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355748, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.753611] env[61970]: DEBUG oslo_concurrency.lockutils [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.854961] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355746, 'name': ReconfigVM_Task, 'duration_secs': 0.339249} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.855253] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 75762ba3-f129-424e-a7cb-962785dfab2a/75762ba3-f129-424e-a7cb-962785dfab2a.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 878.856039] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-499c55ac-2a31-489a-b4af-abad329a27a7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.862980] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 878.862980] env[61970]: value = "task-1355749" [ 878.862980] env[61970]: _type = "Task" [ 878.862980] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.871368] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355749, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.942783] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355747, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.242071} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.943077] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 878.944035] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01214458-3ea9-4359-b3e2-a67a2de0844d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.970128] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 0304640a-6c46-4f57-a8b8-5058da2e3de4/0304640a-6c46-4f57-a8b8-5058da2e3de4.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 878.974023] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a06d8be3-40e2-4b02-a756-75d4b2d97d95 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.988556] env[61970]: INFO nova.compute.manager [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Took 26.08 seconds to build instance. [ 878.994970] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 878.994970] env[61970]: value = "task-1355750" [ 878.994970] env[61970]: _type = "Task" [ 878.994970] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.005181] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355750, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.052093] env[61970]: DEBUG oslo_concurrency.lockutils [None req-958fbf47-fa37-48e8-b698-fdef2789bb2c tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "b19cb516-b163-4bed-ba5b-139a0a18fc05" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.975s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.153858] env[61970]: DEBUG nova.compute.manager [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 879.174691] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355748, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.295470] env[61970]: DEBUG oslo_concurrency.lockutils [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.295796] env[61970]: DEBUG oslo_concurrency.lockutils [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.374515] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355749, 'name': Rename_Task, 'duration_secs': 0.150389} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.374716] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 879.375322] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-09e6b3eb-723e-4250-a045-d527a9781621 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.382251] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 879.382251] env[61970]: value = "task-1355751" [ 879.382251] env[61970]: _type = "Task" [ 879.382251] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.390397] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355751, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.491456] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4289431-d321-44b4-9748-29130846c33b tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "e2185ed0-2bc1-4718-b47d-57150b5e60ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.370s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.507209] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355750, 'name': ReconfigVM_Task, 'duration_secs': 0.301523} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.507982] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 0304640a-6c46-4f57-a8b8-5058da2e3de4/0304640a-6c46-4f57-a8b8-5058da2e3de4.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 879.508256] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a11f531-8ff2-4340-8925-59417074949a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.515279] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 879.515279] env[61970]: value = "task-1355752" [ 879.515279] env[61970]: _type = "Task" [ 879.515279] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.525522] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355752, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.563184] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 5d4d97e1-5812-4884-a1ea-6eb0cee591c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.563418] env[61970]: WARNING nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 6345d239-fbae-4f3a-9f61-79e10539ff0b is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 879.563605] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance b1cc1cc2-15d6-459d-9529-e592ddb225ac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.563773] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance b385407b-1bdd-4c53-907c-cb4c8ce16cc7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.563921] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.564114] env[61970]: WARNING nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance e1415314-f137-4ee3-a065-3d875ef9e7ff is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 879.564293] env[61970]: WARNING nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 035fb797-21c1-4af9-8e66-deee8e9d083a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 879.564443] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 2d280f75-ed84-4db3-9067-8ccc719b0d2e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.564574] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.564735] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 99200b46-7eb7-4ca2-b352-b91b4ad076af actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.564879] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance e2185ed0-2bc1-4718-b47d-57150b5e60ba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.565037] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 75762ba3-f129-424e-a7cb-962785dfab2a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.565210] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 0304640a-6c46-4f57-a8b8-5058da2e3de4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.675790] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355748, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.678959] env[61970]: DEBUG oslo_concurrency.lockutils [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.798824] env[61970]: DEBUG nova.compute.manager [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 879.894970] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355751, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.026709] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355752, 'name': Rename_Task, 'duration_secs': 0.193933} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.026995] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 880.027262] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-623763a2-10b5-4ab5-bf93-a3696b609833 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.032894] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 880.032894] env[61970]: value = "task-1355753" [ 880.032894] env[61970]: _type = "Task" [ 880.032894] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.040827] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355753, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.043362] env[61970]: DEBUG oslo_concurrency.lockutils [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "99200b46-7eb7-4ca2-b352-b91b4ad076af" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.043594] env[61970]: DEBUG oslo_concurrency.lockutils [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "99200b46-7eb7-4ca2-b352-b91b4ad076af" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.043859] env[61970]: DEBUG oslo_concurrency.lockutils [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "99200b46-7eb7-4ca2-b352-b91b4ad076af-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.044249] env[61970]: DEBUG oslo_concurrency.lockutils [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "99200b46-7eb7-4ca2-b352-b91b4ad076af-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.044542] env[61970]: DEBUG oslo_concurrency.lockutils [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "99200b46-7eb7-4ca2-b352-b91b4ad076af-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.046773] env[61970]: INFO nova.compute.manager [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Terminating instance [ 880.048784] env[61970]: DEBUG nova.compute.manager [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 880.048979] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 880.049746] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f54cab7c-fd66-40c2-b9e2-9fcb7b6d6339 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.056267] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 880.056503] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-85ffdc4f-ec64-401f-ae22-84452ba03ac9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.061936] env[61970]: DEBUG oslo_vmware.api [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 880.061936] env[61970]: value = "task-1355754" [ 880.061936] env[61970]: _type = "Task" [ 880.061936] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.067716] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 880.072515] env[61970]: DEBUG oslo_vmware.api [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355754, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.122019] env[61970]: DEBUG oslo_concurrency.lockutils [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "e2185ed0-2bc1-4718-b47d-57150b5e60ba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.122019] env[61970]: DEBUG oslo_concurrency.lockutils [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "e2185ed0-2bc1-4718-b47d-57150b5e60ba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.122019] env[61970]: DEBUG oslo_concurrency.lockutils [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "e2185ed0-2bc1-4718-b47d-57150b5e60ba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.122019] env[61970]: DEBUG oslo_concurrency.lockutils [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "e2185ed0-2bc1-4718-b47d-57150b5e60ba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.122289] env[61970]: DEBUG oslo_concurrency.lockutils [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "e2185ed0-2bc1-4718-b47d-57150b5e60ba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.123634] env[61970]: INFO nova.compute.manager [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Terminating instance [ 880.125775] env[61970]: DEBUG nova.compute.manager [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 880.126649] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 880.128562] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a081c4c-9659-4ad0-a22d-7c3dbeb4454e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.138298] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 880.138298] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f891e48-1b6f-4a48-86e9-8e5587d8f11c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.148891] env[61970]: DEBUG oslo_vmware.api [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 880.148891] env[61970]: value = "task-1355755" [ 880.148891] env[61970]: _type = "Task" [ 880.148891] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.157738] env[61970]: DEBUG oslo_vmware.api [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355755, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.179918] env[61970]: DEBUG oslo_vmware.api [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355748, 'name': RemoveSnapshot_Task, 'duration_secs': 1.310787} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.180787] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Deleted Snapshot of the VM instance {{(pid=61970) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 880.325137] env[61970]: DEBUG oslo_concurrency.lockutils [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.395048] env[61970]: DEBUG oslo_vmware.api [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355751, 'name': PowerOnVM_Task, 'duration_secs': 0.547389} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.395048] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 880.395210] env[61970]: INFO nova.compute.manager [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Took 9.63 seconds to spawn the instance on the hypervisor. [ 880.395444] env[61970]: DEBUG nova.compute.manager [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 880.396242] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd3acc39-5180-48cb-85bf-c654d0acedcc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.543918] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355753, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.571624] env[61970]: DEBUG oslo_vmware.api [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355754, 'name': PowerOffVM_Task, 'duration_secs': 0.265487} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.571912] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 880.572090] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 880.572349] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8a50fc63-1811-4d11-8783-c97e997a585d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.574492] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance d0ad9689-bdb7-4d68-b19a-3f805ab4612e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 880.574733] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 880.575082] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2432MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 880.658522] env[61970]: DEBUG oslo_vmware.api [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355755, 'name': PowerOffVM_Task, 'duration_secs': 0.184996} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.658802] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 880.658969] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 880.659241] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bb5cbe2a-15f3-4c0d-9037-10e31a1d061d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.691606] env[61970]: WARNING nova.compute.manager [None req-f5d12470-b7ef-4a12-8108-d8ae9d1661e4 tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Image not found during snapshot: nova.exception.ImageNotFound: Image 34020495-c8a7-4322-81b7-a1494c1485f7 could not be found. [ 880.724463] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 880.724823] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 880.725263] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Deleting the datastore file [datastore1] e2185ed0-2bc1-4718-b47d-57150b5e60ba {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 880.725588] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-43bb9774-a581-4d8a-bc2d-f50af50a79a0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.734343] env[61970]: DEBUG oslo_vmware.api [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 880.734343] env[61970]: value = "task-1355758" [ 880.734343] env[61970]: _type = "Task" [ 880.734343] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.743664] env[61970]: DEBUG oslo_vmware.api [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355758, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.787978] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1627110c-ed03-4801-9ce0-d04216f9ea75 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.795072] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-352304bd-ffaa-4d4a-9d88-2d55cf4a3948 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.828586] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71774c0-fcb0-41bb-b9af-b33036ece158 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.837693] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37517a7d-18ef-4792-996d-c7af7f561b64 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.852436] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.912937] env[61970]: INFO nova.compute.manager [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Took 27.58 seconds to build instance. [ 881.044599] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355753, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.245205] env[61970]: DEBUG oslo_vmware.api [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355758, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.294764] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 881.294764] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 881.294764] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Deleting the datastore file [datastore2] 99200b46-7eb7-4ca2-b352-b91b4ad076af {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 881.294764] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1fd382b1-0176-4678-9469-d5e788109880 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.300193] env[61970]: DEBUG oslo_vmware.api [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for the task: (returnval){ [ 881.300193] env[61970]: value = "task-1355759" [ 881.300193] env[61970]: _type = "Task" [ 881.300193] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.307441] env[61970]: DEBUG oslo_vmware.api [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355759, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.324923] env[61970]: DEBUG oslo_concurrency.lockutils [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "0c50937e-4646-47ec-ac0c-0281c9424d86" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.325159] env[61970]: DEBUG oslo_concurrency.lockutils [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "0c50937e-4646-47ec-ac0c-0281c9424d86" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.355487] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 881.415173] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99b92c9a-369c-4bc6-b0c8-acfbf1bdb713 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "75762ba3-f129-424e-a7cb-962785dfab2a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.860s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.546273] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355753, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.557029] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "2d280f75-ed84-4db3-9067-8ccc719b0d2e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.557127] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "2d280f75-ed84-4db3-9067-8ccc719b0d2e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.557266] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "2d280f75-ed84-4db3-9067-8ccc719b0d2e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.557448] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "2d280f75-ed84-4db3-9067-8ccc719b0d2e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.557619] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "2d280f75-ed84-4db3-9067-8ccc719b0d2e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.559763] env[61970]: INFO nova.compute.manager [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Terminating instance [ 881.562047] env[61970]: DEBUG nova.compute.manager [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 881.562047] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 881.562465] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a237cd-739f-41d8-ad53-fa9b43d2474e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.572728] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 881.572728] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-153ae24a-6bd2-4abf-bc7b-e28b0d380594 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.578747] env[61970]: DEBUG oslo_vmware.api [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for the task: (returnval){ [ 881.578747] env[61970]: value = "task-1355760" [ 881.578747] env[61970]: _type = "Task" [ 881.578747] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.586882] env[61970]: DEBUG oslo_vmware.api [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355760, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.744707] env[61970]: DEBUG oslo_vmware.api [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355758, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.810045] env[61970]: DEBUG oslo_vmware.api [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355759, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.444925} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.810350] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 881.810556] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 881.810734] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 881.810905] env[61970]: INFO nova.compute.manager [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Took 1.76 seconds to destroy the instance on the hypervisor. [ 881.811249] env[61970]: DEBUG oslo.service.loopingcall [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.811464] env[61970]: DEBUG nova.compute.manager [-] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 881.811561] env[61970]: DEBUG nova.network.neutron [-] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 881.827493] env[61970]: DEBUG nova.compute.manager [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 881.859623] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 881.859789] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.341s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.860051] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.381s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.860242] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.862137] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 8.932s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.862324] env[61970]: DEBUG nova.objects.instance [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61970) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 881.889379] env[61970]: INFO nova.scheduler.client.report [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Deleted allocations for instance 035fb797-21c1-4af9-8e66-deee8e9d083a [ 882.045967] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355753, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.093828] env[61970]: DEBUG oslo_vmware.api [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355760, 'name': PowerOffVM_Task, 'duration_secs': 0.177398} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.095851] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 882.096064] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 882.096334] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-edc855c9-0f07-49ab-bb22-fd76a0f7a847 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.107197] env[61970]: DEBUG nova.compute.manager [req-dfa464c0-23ba-47ee-bc31-69fffd3d5a6f req-78f2b554-756c-41d4-afe0-113aeb48ec8c service nova] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Received event network-vif-deleted-f0126a7e-adec-4042-8b0e-5eb42bbd1572 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 882.107197] env[61970]: INFO nova.compute.manager [req-dfa464c0-23ba-47ee-bc31-69fffd3d5a6f req-78f2b554-756c-41d4-afe0-113aeb48ec8c service nova] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Neutron deleted interface f0126a7e-adec-4042-8b0e-5eb42bbd1572; detaching it from the instance and deleting it from the info cache [ 882.107197] env[61970]: DEBUG nova.network.neutron [req-dfa464c0-23ba-47ee-bc31-69fffd3d5a6f req-78f2b554-756c-41d4-afe0-113aeb48ec8c service nova] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.159896] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 882.159896] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 882.160075] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Deleting the datastore file [datastore2] 2d280f75-ed84-4db3-9067-8ccc719b0d2e {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 882.160355] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9532f6cf-0c62-403a-ad98-84b1226eea02 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.166297] env[61970]: DEBUG oslo_vmware.api [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for the task: (returnval){ [ 882.166297] env[61970]: value = "task-1355762" [ 882.166297] env[61970]: _type = "Task" [ 882.166297] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.173836] env[61970]: DEBUG oslo_vmware.api [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355762, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.246598] env[61970]: DEBUG oslo_vmware.api [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355758, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.349787] env[61970]: DEBUG oslo_concurrency.lockutils [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.398831] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3ecffe4f-45ea-4261-9965-23ce1971cefc tempest-InstanceActionsNegativeTestJSON-1024516290 tempest-InstanceActionsNegativeTestJSON-1024516290-project-member] Lock "035fb797-21c1-4af9-8e66-deee8e9d083a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.854s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.546696] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355753, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.586654] env[61970]: DEBUG nova.network.neutron [-] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.609974] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-385834dc-d441-4d96-96eb-e9ad22da837b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.621703] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc325fed-5c7d-4761-8f57-68afff2951db {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.647899] env[61970]: DEBUG nova.compute.manager [req-dfa464c0-23ba-47ee-bc31-69fffd3d5a6f req-78f2b554-756c-41d4-afe0-113aeb48ec8c service nova] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Detach interface failed, port_id=f0126a7e-adec-4042-8b0e-5eb42bbd1572, reason: Instance 99200b46-7eb7-4ca2-b352-b91b4ad076af could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 882.683990] env[61970]: DEBUG oslo_vmware.api [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Task: {'id': task-1355762, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145522} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.684289] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 882.684474] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 882.684693] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 882.684873] env[61970]: INFO nova.compute.manager [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Took 1.12 seconds to destroy the instance on the hypervisor. [ 882.685129] env[61970]: DEBUG oslo.service.loopingcall [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.685345] env[61970]: DEBUG nova.compute.manager [-] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 882.685550] env[61970]: DEBUG nova.network.neutron [-] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 882.747343] env[61970]: DEBUG oslo_vmware.api [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355758, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.870889] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c26c8d6-9dfc-4313-a35a-5c7df13831b4 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.872131] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.984s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.872342] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.875148] env[61970]: DEBUG oslo_concurrency.lockutils [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.122s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.875350] env[61970]: DEBUG oslo_concurrency.lockutils [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.877292] env[61970]: DEBUG oslo_concurrency.lockutils [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.198s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.879266] env[61970]: INFO nova.compute.claims [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.903222] env[61970]: INFO nova.scheduler.client.report [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Deleted allocations for instance e1415314-f137-4ee3-a065-3d875ef9e7ff [ 882.910490] env[61970]: INFO nova.scheduler.client.report [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Deleted allocations for instance 6345d239-fbae-4f3a-9f61-79e10539ff0b [ 883.052320] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355753, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.089161] env[61970]: INFO nova.compute.manager [-] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Took 1.28 seconds to deallocate network for instance. [ 883.250330] env[61970]: DEBUG oslo_vmware.api [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Task: {'id': task-1355758, 'name': DeleteDatastoreFile_Task, 'duration_secs': 2.160336} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.250601] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 883.250816] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 883.250988] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 883.251171] env[61970]: INFO nova.compute.manager [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Took 3.13 seconds to destroy the instance on the hypervisor. [ 883.251427] env[61970]: DEBUG oslo.service.loopingcall [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.251630] env[61970]: DEBUG nova.compute.manager [-] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 883.251719] env[61970]: DEBUG nova.network.neutron [-] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 883.414163] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3f02e68a-a4de-4fe6-905d-d554a6c20343 tempest-ServerShowV257Test-957820058 tempest-ServerShowV257Test-957820058-project-member] Lock "e1415314-f137-4ee3-a065-3d875ef9e7ff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.326s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.427486] env[61970]: DEBUG oslo_concurrency.lockutils [None req-91d76e68-065a-4a53-b7dd-98afe112148f tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "6345d239-fbae-4f3a-9f61-79e10539ff0b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.688s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.436333] env[61970]: DEBUG nova.compute.manager [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Stashing vm_state: active {{(pid=61970) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 883.548656] env[61970]: DEBUG oslo_vmware.api [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355753, 'name': PowerOnVM_Task, 'duration_secs': 3.141922} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.549209] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 883.549615] env[61970]: INFO nova.compute.manager [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Took 10.39 seconds to spawn the instance on the hypervisor. [ 883.549886] env[61970]: DEBUG nova.compute.manager [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 883.550729] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c5c326-7363-4a69-ba89-e0745cc89c19 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.585896] env[61970]: DEBUG nova.network.neutron [-] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.596206] env[61970]: DEBUG oslo_concurrency.lockutils [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.701693] env[61970]: DEBUG oslo_concurrency.lockutils [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "5d4d97e1-5812-4884-a1ea-6eb0cee591c0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.702237] env[61970]: DEBUG oslo_concurrency.lockutils [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "5d4d97e1-5812-4884-a1ea-6eb0cee591c0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.702463] env[61970]: DEBUG oslo_concurrency.lockutils [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "5d4d97e1-5812-4884-a1ea-6eb0cee591c0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.702662] env[61970]: DEBUG oslo_concurrency.lockutils [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "5d4d97e1-5812-4884-a1ea-6eb0cee591c0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.702839] env[61970]: DEBUG oslo_concurrency.lockutils [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "5d4d97e1-5812-4884-a1ea-6eb0cee591c0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.705017] env[61970]: INFO nova.compute.manager [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Terminating instance [ 883.706806] env[61970]: DEBUG nova.compute.manager [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 883.706992] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 883.707918] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5133b53b-83fb-4d7f-a8b3-62dbdcf2a6bd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.715827] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 883.716091] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-87ac73fd-9a0b-48ab-a556-e31ab188d1dd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.722841] env[61970]: DEBUG oslo_vmware.api [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 883.722841] env[61970]: value = "task-1355763" [ 883.722841] env[61970]: _type = "Task" [ 883.722841] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.730665] env[61970]: DEBUG oslo_vmware.api [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355763, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.970379] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.073731] env[61970]: INFO nova.compute.manager [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Took 29.23 seconds to build instance. [ 884.089550] env[61970]: INFO nova.compute.manager [-] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Took 1.40 seconds to deallocate network for instance. [ 884.091523] env[61970]: DEBUG nova.network.neutron [-] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.183142] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afc687a7-3577-431a-aef0-8e5aeb9fb2bc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.191843] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-607c9a0e-ab1c-4a43-bc7b-03f3f94381ec {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.237994] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5873ac3a-5438-4c8a-8157-d291c10dfd28 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.249948] env[61970]: DEBUG oslo_vmware.api [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355763, 'name': PowerOffVM_Task, 'duration_secs': 0.28622} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.251836] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f3551a-2c72-49ed-be7f-3b10e314de63 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.258237] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 884.258237] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 884.258237] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3b637922-5aa7-40b6-9ca3-15309d9b7b8f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.269218] env[61970]: DEBUG nova.compute.provider_tree [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.327686] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 884.327922] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 884.327922] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Deleting the datastore file [datastore2] 5d4d97e1-5812-4884-a1ea-6eb0cee591c0 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 884.329428] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c524143e-c31a-4e1a-8042-67e8a7d47daf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.334818] env[61970]: DEBUG oslo_vmware.api [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for the task: (returnval){ [ 884.334818] env[61970]: value = "task-1355765" [ 884.334818] env[61970]: _type = "Task" [ 884.334818] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.343637] env[61970]: DEBUG oslo_vmware.api [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355765, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.363334] env[61970]: DEBUG nova.compute.manager [req-e38dcfa8-c942-4e19-90b4-e84d7f609f14 req-2ef78375-2c08-4c24-aca5-a9d342750037 service nova] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Received event network-vif-deleted-fbc5f4e5-3299-4e59-90aa-badfe6a2ca4c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.363575] env[61970]: DEBUG nova.compute.manager [req-e38dcfa8-c942-4e19-90b4-e84d7f609f14 req-2ef78375-2c08-4c24-aca5-a9d342750037 service nova] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Received event network-vif-deleted-3cec35ca-1e33-4861-8a90-71573ae58d5c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.575243] env[61970]: DEBUG oslo_concurrency.lockutils [None req-09e382d8-e8ff-485f-a19b-7b689c444776 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "0304640a-6c46-4f57-a8b8-5058da2e3de4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.806s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.599091] env[61970]: INFO nova.compute.manager [-] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Took 1.35 seconds to deallocate network for instance. [ 884.599824] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.658095] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b46d4152-36e9-4c96-ba3b-775ef268856c tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "interface-b1cc1cc2-15d6-459d-9529-e592ddb225ac-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.658389] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b46d4152-36e9-4c96-ba3b-775ef268856c tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "interface-b1cc1cc2-15d6-459d-9529-e592ddb225ac-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.658695] env[61970]: DEBUG nova.objects.instance [None req-b46d4152-36e9-4c96-ba3b-775ef268856c tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lazy-loading 'flavor' on Instance uuid b1cc1cc2-15d6-459d-9529-e592ddb225ac {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 884.774441] env[61970]: DEBUG nova.scheduler.client.report [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.847150] env[61970]: DEBUG oslo_vmware.api [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355765, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.106946] env[61970]: DEBUG oslo_concurrency.lockutils [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.162900] env[61970]: DEBUG nova.objects.instance [None req-b46d4152-36e9-4c96-ba3b-775ef268856c tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lazy-loading 'pci_requests' on Instance uuid b1cc1cc2-15d6-459d-9529-e592ddb225ac {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.280462] env[61970]: DEBUG oslo_concurrency.lockutils [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.403s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.280963] env[61970]: DEBUG nova.compute.manager [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 885.283979] env[61970]: DEBUG oslo_concurrency.lockutils [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.959s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.288511] env[61970]: INFO nova.compute.claims [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 885.349529] env[61970]: DEBUG oslo_vmware.api [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Task: {'id': task-1355765, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.579848} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.349860] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 885.350699] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 885.350699] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 885.350699] env[61970]: INFO nova.compute.manager [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Took 1.64 seconds to destroy the instance on the hypervisor. [ 885.350699] env[61970]: DEBUG oslo.service.loopingcall [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.350864] env[61970]: DEBUG nova.compute.manager [-] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 885.350864] env[61970]: DEBUG nova.network.neutron [-] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 885.665542] env[61970]: DEBUG nova.objects.base [None req-b46d4152-36e9-4c96-ba3b-775ef268856c tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 885.665807] env[61970]: DEBUG nova.network.neutron [None req-b46d4152-36e9-4c96-ba3b-775ef268856c tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 885.797488] env[61970]: DEBUG nova.compute.utils [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.803624] env[61970]: DEBUG nova.compute.manager [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 885.803852] env[61970]: DEBUG nova.network.neutron [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 885.844573] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b46d4152-36e9-4c96-ba3b-775ef268856c tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "interface-b1cc1cc2-15d6-459d-9529-e592ddb225ac-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.186s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.875401] env[61970]: DEBUG nova.policy [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64082edf5f2e4621b5f3586e1792e42a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9235eeeb6fa4ccf959ffbea456f3694', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 886.129884] env[61970]: DEBUG nova.network.neutron [-] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.306457] env[61970]: DEBUG nova.compute.manager [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 886.364890] env[61970]: DEBUG nova.network.neutron [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Successfully created port: 7d01e768-116a-4e5c-b85a-8a9920781421 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.417111] env[61970]: DEBUG nova.compute.manager [req-96088bc2-8704-4ad0-992a-a9dabe83d2ee req-4000cd89-6514-4051-ba16-b430b07ec452 service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Received event network-changed-2545e2b0-a432-44ae-a287-0b68ceab7031 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.417111] env[61970]: DEBUG nova.compute.manager [req-96088bc2-8704-4ad0-992a-a9dabe83d2ee req-4000cd89-6514-4051-ba16-b430b07ec452 service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Refreshing instance network info cache due to event network-changed-2545e2b0-a432-44ae-a287-0b68ceab7031. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 886.417111] env[61970]: DEBUG oslo_concurrency.lockutils [req-96088bc2-8704-4ad0-992a-a9dabe83d2ee req-4000cd89-6514-4051-ba16-b430b07ec452 service nova] Acquiring lock "refresh_cache-0304640a-6c46-4f57-a8b8-5058da2e3de4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.417111] env[61970]: DEBUG oslo_concurrency.lockutils [req-96088bc2-8704-4ad0-992a-a9dabe83d2ee req-4000cd89-6514-4051-ba16-b430b07ec452 service nova] Acquired lock "refresh_cache-0304640a-6c46-4f57-a8b8-5058da2e3de4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.417111] env[61970]: DEBUG nova.network.neutron [req-96088bc2-8704-4ad0-992a-a9dabe83d2ee req-4000cd89-6514-4051-ba16-b430b07ec452 service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Refreshing network info cache for port 2545e2b0-a432-44ae-a287-0b68ceab7031 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 886.547602] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2feb17e8-9a0d-4dea-8382-5a6c292835c5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.557075] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e79de7d-7c03-4859-907a-ecf88fee256f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.595509] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d150c48-6d6b-4cd8-9216-4b0282ca0aa9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.603962] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0857eb38-35de-428e-9d3c-fd0fcc8bd121 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.620261] env[61970]: DEBUG nova.compute.provider_tree [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 886.632158] env[61970]: INFO nova.compute.manager [-] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Took 1.28 seconds to deallocate network for instance. [ 887.138634] env[61970]: DEBUG oslo_concurrency.lockutils [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.146677] env[61970]: DEBUG nova.network.neutron [req-96088bc2-8704-4ad0-992a-a9dabe83d2ee req-4000cd89-6514-4051-ba16-b430b07ec452 service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Updated VIF entry in instance network info cache for port 2545e2b0-a432-44ae-a287-0b68ceab7031. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 887.147087] env[61970]: DEBUG nova.network.neutron [req-96088bc2-8704-4ad0-992a-a9dabe83d2ee req-4000cd89-6514-4051-ba16-b430b07ec452 service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Updating instance_info_cache with network_info: [{"id": "2545e2b0-a432-44ae-a287-0b68ceab7031", "address": "fa:16:3e:a6:77:ea", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2545e2b0-a4", "ovs_interfaceid": "2545e2b0-a432-44ae-a287-0b68ceab7031", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.148840] env[61970]: ERROR nova.scheduler.client.report [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [req-e9985bf3-fdf3-4a36-a2f6-16cd6602efb3] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fc0f3e4e-5ef5-45e9-8335-cc68743405aa. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e9985bf3-fdf3-4a36-a2f6-16cd6602efb3"}]} [ 887.169458] env[61970]: DEBUG nova.scheduler.client.report [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Refreshing inventories for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 887.188127] env[61970]: DEBUG nova.scheduler.client.report [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Updating ProviderTree inventory for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 887.188366] env[61970]: DEBUG nova.compute.provider_tree [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 887.207659] env[61970]: DEBUG nova.scheduler.client.report [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Refreshing aggregate associations for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa, aggregates: None {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 887.226700] env[61970]: DEBUG nova.scheduler.client.report [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Refreshing trait associations for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 887.324745] env[61970]: DEBUG nova.compute.manager [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 887.360758] env[61970]: DEBUG nova.virt.hardware [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 887.361065] env[61970]: DEBUG nova.virt.hardware [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 887.361211] env[61970]: DEBUG nova.virt.hardware [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.361560] env[61970]: DEBUG nova.virt.hardware [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 887.361745] env[61970]: DEBUG nova.virt.hardware [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.361891] env[61970]: DEBUG nova.virt.hardware [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 887.362110] env[61970]: DEBUG nova.virt.hardware [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 887.362432] env[61970]: DEBUG nova.virt.hardware [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 887.362648] env[61970]: DEBUG nova.virt.hardware [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 887.362860] env[61970]: DEBUG nova.virt.hardware [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 887.363101] env[61970]: DEBUG nova.virt.hardware [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 887.364321] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f369f7-d557-4c2a-a873-701ad29f07cb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.379614] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-095bbc4a-4458-4d32-9a31-f96e2472aea7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.461818] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a50390-962e-49c0-985d-e847acdf8834 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.469519] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a27be3d-cd66-40ca-b601-88b4f2b69356 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.500949] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc2439d-ea70-4183-a3ae-54877c5270c1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.510075] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-494d7ebf-fb6b-46fd-9512-b10e85d11688 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.523733] env[61970]: DEBUG nova.compute.provider_tree [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 887.651786] env[61970]: DEBUG oslo_concurrency.lockutils [req-96088bc2-8704-4ad0-992a-a9dabe83d2ee req-4000cd89-6514-4051-ba16-b430b07ec452 service nova] Releasing lock "refresh_cache-0304640a-6c46-4f57-a8b8-5058da2e3de4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.652080] env[61970]: DEBUG nova.compute.manager [req-96088bc2-8704-4ad0-992a-a9dabe83d2ee req-4000cd89-6514-4051-ba16-b430b07ec452 service nova] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Received event network-vif-deleted-1080e2f3-c9a3-48bc-bd65-b5f231bacdc7 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.707678] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "interface-b1cc1cc2-15d6-459d-9529-e592ddb225ac-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.707996] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "interface-b1cc1cc2-15d6-459d-9529-e592ddb225ac-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.708314] env[61970]: DEBUG nova.objects.instance [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lazy-loading 'flavor' on Instance uuid b1cc1cc2-15d6-459d-9529-e592ddb225ac {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 888.075066] env[61970]: DEBUG nova.scheduler.client.report [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Updated inventory for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with generation 93 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 888.075332] env[61970]: DEBUG nova.compute.provider_tree [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Updating resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa generation from 93 to 94 during operation: update_inventory {{(pid=61970) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 888.076562] env[61970]: DEBUG nova.compute.provider_tree [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 888.208491] env[61970]: DEBUG nova.compute.manager [req-9950286c-b472-4e75-a9fb-f65111a5646d req-fdd6895e-a16b-4714-804a-8057e0befeb3 service nova] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Received event network-vif-plugged-7d01e768-116a-4e5c-b85a-8a9920781421 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.208796] env[61970]: DEBUG oslo_concurrency.lockutils [req-9950286c-b472-4e75-a9fb-f65111a5646d req-fdd6895e-a16b-4714-804a-8057e0befeb3 service nova] Acquiring lock "8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.209018] env[61970]: DEBUG oslo_concurrency.lockutils [req-9950286c-b472-4e75-a9fb-f65111a5646d req-fdd6895e-a16b-4714-804a-8057e0befeb3 service nova] Lock "8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.209161] env[61970]: DEBUG oslo_concurrency.lockutils [req-9950286c-b472-4e75-a9fb-f65111a5646d req-fdd6895e-a16b-4714-804a-8057e0befeb3 service nova] Lock "8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.209388] env[61970]: DEBUG nova.compute.manager [req-9950286c-b472-4e75-a9fb-f65111a5646d req-fdd6895e-a16b-4714-804a-8057e0befeb3 service nova] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] No waiting events found dispatching network-vif-plugged-7d01e768-116a-4e5c-b85a-8a9920781421 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 888.209564] env[61970]: WARNING nova.compute.manager [req-9950286c-b472-4e75-a9fb-f65111a5646d req-fdd6895e-a16b-4714-804a-8057e0befeb3 service nova] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Received unexpected event network-vif-plugged-7d01e768-116a-4e5c-b85a-8a9920781421 for instance with vm_state building and task_state spawning. [ 888.250234] env[61970]: DEBUG nova.network.neutron [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Successfully updated port: 7d01e768-116a-4e5c-b85a-8a9920781421 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 888.324835] env[61970]: DEBUG nova.objects.instance [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lazy-loading 'pci_requests' on Instance uuid b1cc1cc2-15d6-459d-9529-e592ddb225ac {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 888.583843] env[61970]: DEBUG oslo_concurrency.lockutils [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.300s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.584387] env[61970]: DEBUG nova.compute.manager [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 888.587113] env[61970]: DEBUG oslo_concurrency.lockutils [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.238s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.588532] env[61970]: INFO nova.compute.claims [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 888.755273] env[61970]: DEBUG oslo_concurrency.lockutils [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "refresh_cache-8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.755273] env[61970]: DEBUG oslo_concurrency.lockutils [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "refresh_cache-8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.755273] env[61970]: DEBUG nova.network.neutron [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.828356] env[61970]: DEBUG nova.objects.base [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 888.828588] env[61970]: DEBUG nova.network.neutron [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 888.878831] env[61970]: DEBUG nova.policy [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2af915dc43c84940a937dfb8a04ea0b6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17ca05a1e6664430a02de563d98c1148', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 889.092828] env[61970]: DEBUG nova.compute.utils [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 889.096165] env[61970]: DEBUG nova.compute.manager [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 889.096339] env[61970]: DEBUG nova.network.neutron [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 889.140224] env[61970]: DEBUG nova.policy [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb2ab679e39540f09ce1fa95971677f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a44c05a894394d69b8b4fd36c7b2532a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 889.213988] env[61970]: DEBUG nova.network.neutron [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Successfully created port: eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 889.322302] env[61970]: DEBUG nova.network.neutron [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 889.427117] env[61970]: DEBUG nova.network.neutron [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Successfully created port: b83024be-4171-40cd-9d1c-65a2d2ce8b9e {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 889.564646] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "7f59000a-94d9-45b6-aa7a-300d95793615" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.564885] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "7f59000a-94d9-45b6-aa7a-300d95793615" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.597271] env[61970]: DEBUG nova.compute.manager [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 889.613092] env[61970]: DEBUG nova.network.neutron [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Updating instance_info_cache with network_info: [{"id": "7d01e768-116a-4e5c-b85a-8a9920781421", "address": "fa:16:3e:6a:24:45", "network": {"id": "e02bfa6c-8f0d-4413-8dcf-e644c07792d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1410864732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9235eeeb6fa4ccf959ffbea456f3694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d01e768-11", "ovs_interfaceid": "7d01e768-116a-4e5c-b85a-8a9920781421", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.809587] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f515342-84ae-46ce-98a1-60acefa69e85 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.817395] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b961884-237b-4496-8820-d50b73a6e7ba {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.846531] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17567f5-3876-41e6-8aa7-633096133cf8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.853731] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d38ee5-eaa4-4728-83d5-651a8cc4b360 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.866512] env[61970]: DEBUG nova.compute.provider_tree [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.071330] env[61970]: DEBUG nova.compute.manager [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 890.117280] env[61970]: DEBUG oslo_concurrency.lockutils [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "refresh_cache-8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.117571] env[61970]: DEBUG nova.compute.manager [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Instance network_info: |[{"id": "7d01e768-116a-4e5c-b85a-8a9920781421", "address": "fa:16:3e:6a:24:45", "network": {"id": "e02bfa6c-8f0d-4413-8dcf-e644c07792d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1410864732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9235eeeb6fa4ccf959ffbea456f3694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d01e768-11", "ovs_interfaceid": "7d01e768-116a-4e5c-b85a-8a9920781421", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 890.118993] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:24:45', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '510d3c47-3615-43d5-aa5d-a279fd915e71', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7d01e768-116a-4e5c-b85a-8a9920781421', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 890.125659] env[61970]: DEBUG oslo.service.loopingcall [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.126328] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 890.126565] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e7b0991a-f489-4678-aec2-5f3d8bc98547 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.147670] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.147670] env[61970]: value = "task-1355766" [ 890.147670] env[61970]: _type = "Task" [ 890.147670] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.155710] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355766, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.274280] env[61970]: DEBUG nova.compute.manager [req-f7ec77e7-6347-4ff2-83ce-c051bf3f95a3 req-2e1be6b4-cf89-4905-a094-c33005acfad2 service nova] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Received event network-changed-7d01e768-116a-4e5c-b85a-8a9920781421 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.274551] env[61970]: DEBUG nova.compute.manager [req-f7ec77e7-6347-4ff2-83ce-c051bf3f95a3 req-2e1be6b4-cf89-4905-a094-c33005acfad2 service nova] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Refreshing instance network info cache due to event network-changed-7d01e768-116a-4e5c-b85a-8a9920781421. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 890.274600] env[61970]: DEBUG oslo_concurrency.lockutils [req-f7ec77e7-6347-4ff2-83ce-c051bf3f95a3 req-2e1be6b4-cf89-4905-a094-c33005acfad2 service nova] Acquiring lock "refresh_cache-8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.274739] env[61970]: DEBUG oslo_concurrency.lockutils [req-f7ec77e7-6347-4ff2-83ce-c051bf3f95a3 req-2e1be6b4-cf89-4905-a094-c33005acfad2 service nova] Acquired lock "refresh_cache-8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.274896] env[61970]: DEBUG nova.network.neutron [req-f7ec77e7-6347-4ff2-83ce-c051bf3f95a3 req-2e1be6b4-cf89-4905-a094-c33005acfad2 service nova] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Refreshing network info cache for port 7d01e768-116a-4e5c-b85a-8a9920781421 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 890.369656] env[61970]: DEBUG nova.scheduler.client.report [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.590613] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.610902] env[61970]: DEBUG nova.compute.manager [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 890.635246] env[61970]: DEBUG nova.virt.hardware [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 890.635550] env[61970]: DEBUG nova.virt.hardware [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 890.635710] env[61970]: DEBUG nova.virt.hardware [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 890.635928] env[61970]: DEBUG nova.virt.hardware [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 890.636123] env[61970]: DEBUG nova.virt.hardware [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 890.636292] env[61970]: DEBUG nova.virt.hardware [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 890.636495] env[61970]: DEBUG nova.virt.hardware [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 890.636649] env[61970]: DEBUG nova.virt.hardware [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 890.636811] env[61970]: DEBUG nova.virt.hardware [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 890.636974] env[61970]: DEBUG nova.virt.hardware [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 890.637174] env[61970]: DEBUG nova.virt.hardware [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 890.638032] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f3804af-1c04-4289-9e2c-83e343cf60ec {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.646073] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a08b271-a77b-4f93-b062-cbc9f4b590c9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.664032] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355766, 'name': CreateVM_Task, 'duration_secs': 0.318606} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.664183] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 890.664799] env[61970]: DEBUG oslo_concurrency.lockutils [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.665225] env[61970]: DEBUG oslo_concurrency.lockutils [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.665273] env[61970]: DEBUG oslo_concurrency.lockutils [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 890.665481] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d852c919-58b0-4794-8f44-9375522411ed {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.669539] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 890.669539] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e823f7-480a-bfae-c9e7-6e44f25254ab" [ 890.669539] env[61970]: _type = "Task" [ 890.669539] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.676453] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e823f7-480a-bfae-c9e7-6e44f25254ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.877658] env[61970]: DEBUG oslo_concurrency.lockutils [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.290s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.878224] env[61970]: DEBUG nova.compute.manager [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 890.886948] env[61970]: DEBUG oslo_concurrency.lockutils [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.291s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.891790] env[61970]: DEBUG nova.objects.instance [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lazy-loading 'resources' on Instance uuid 99200b46-7eb7-4ca2-b352-b91b4ad076af {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 890.984900] env[61970]: DEBUG nova.network.neutron [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Successfully updated port: b83024be-4171-40cd-9d1c-65a2d2ce8b9e {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 891.123846] env[61970]: DEBUG nova.compute.manager [req-09705595-d94c-4840-a8dc-3f3c97bc4563 req-d1ff5066-0979-4a8a-a6b7-544dede4241e service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Received event network-vif-plugged-eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 891.124124] env[61970]: DEBUG oslo_concurrency.lockutils [req-09705595-d94c-4840-a8dc-3f3c97bc4563 req-d1ff5066-0979-4a8a-a6b7-544dede4241e service nova] Acquiring lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.124339] env[61970]: DEBUG oslo_concurrency.lockutils [req-09705595-d94c-4840-a8dc-3f3c97bc4563 req-d1ff5066-0979-4a8a-a6b7-544dede4241e service nova] Lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.124502] env[61970]: DEBUG oslo_concurrency.lockutils [req-09705595-d94c-4840-a8dc-3f3c97bc4563 req-d1ff5066-0979-4a8a-a6b7-544dede4241e service nova] Lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.124672] env[61970]: DEBUG nova.compute.manager [req-09705595-d94c-4840-a8dc-3f3c97bc4563 req-d1ff5066-0979-4a8a-a6b7-544dede4241e service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] No waiting events found dispatching network-vif-plugged-eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 891.124835] env[61970]: WARNING nova.compute.manager [req-09705595-d94c-4840-a8dc-3f3c97bc4563 req-d1ff5066-0979-4a8a-a6b7-544dede4241e service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Received unexpected event network-vif-plugged-eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9 for instance with vm_state active and task_state None. [ 891.181966] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e823f7-480a-bfae-c9e7-6e44f25254ab, 'name': SearchDatastore_Task, 'duration_secs': 0.009006} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.182179] env[61970]: DEBUG oslo_concurrency.lockutils [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.182408] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.182630] env[61970]: DEBUG oslo_concurrency.lockutils [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.182773] env[61970]: DEBUG oslo_concurrency.lockutils [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.182946] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.183218] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62051159-c360-4c15-b51f-15b515e90d6f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.191825] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.192074] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 891.192814] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8ad27a9-4e67-447d-8c75-0e989509a3b3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.198122] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 891.198122] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529f1b33-453b-2f32-9660-93cbfe772539" [ 891.198122] env[61970]: _type = "Task" [ 891.198122] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.206529] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529f1b33-453b-2f32-9660-93cbfe772539, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.230207] env[61970]: DEBUG nova.network.neutron [req-f7ec77e7-6347-4ff2-83ce-c051bf3f95a3 req-2e1be6b4-cf89-4905-a094-c33005acfad2 service nova] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Updated VIF entry in instance network info cache for port 7d01e768-116a-4e5c-b85a-8a9920781421. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 891.230594] env[61970]: DEBUG nova.network.neutron [req-f7ec77e7-6347-4ff2-83ce-c051bf3f95a3 req-2e1be6b4-cf89-4905-a094-c33005acfad2 service nova] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Updating instance_info_cache with network_info: [{"id": "7d01e768-116a-4e5c-b85a-8a9920781421", "address": "fa:16:3e:6a:24:45", "network": {"id": "e02bfa6c-8f0d-4413-8dcf-e644c07792d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1410864732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9235eeeb6fa4ccf959ffbea456f3694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d01e768-11", "ovs_interfaceid": "7d01e768-116a-4e5c-b85a-8a9920781421", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.306020] env[61970]: DEBUG nova.network.neutron [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Successfully updated port: eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 891.386079] env[61970]: DEBUG nova.compute.utils [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 891.391956] env[61970]: DEBUG nova.compute.manager [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 891.392198] env[61970]: DEBUG nova.network.neutron [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 891.442415] env[61970]: DEBUG nova.policy [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd04344afddb14418b5018b69fbd7694e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c6a4997482a4834a30c79c4cbab2cc1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 891.488110] env[61970]: DEBUG oslo_concurrency.lockutils [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "refresh_cache-d0ad9689-bdb7-4d68-b19a-3f805ab4612e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.488285] env[61970]: DEBUG oslo_concurrency.lockutils [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquired lock "refresh_cache-d0ad9689-bdb7-4d68-b19a-3f805ab4612e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.488435] env[61970]: DEBUG nova.network.neutron [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 891.611704] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c0638a-2087-438b-8941-a7c9c9115027 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.619277] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ff04b91-eac9-4843-83e5-f36440d83b39 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.649983] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4975db8e-cf26-4ce4-a468-b93d31431731 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.658988] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9756c32e-0a87-41c8-bd1b-e05e358d0cf0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.675610] env[61970]: DEBUG nova.compute.provider_tree [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 891.709246] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529f1b33-453b-2f32-9660-93cbfe772539, 'name': SearchDatastore_Task, 'duration_secs': 0.009516} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.710050] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e11bba72-f887-4bad-92ff-cd84875fbb37 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.715461] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 891.715461] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]524628ba-6b24-52dc-c9a6-c60d5ff8d351" [ 891.715461] env[61970]: _type = "Task" [ 891.715461] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.723896] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]524628ba-6b24-52dc-c9a6-c60d5ff8d351, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.733621] env[61970]: DEBUG oslo_concurrency.lockutils [req-f7ec77e7-6347-4ff2-83ce-c051bf3f95a3 req-2e1be6b4-cf89-4905-a094-c33005acfad2 service nova] Releasing lock "refresh_cache-8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.804417] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.804604] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.804783] env[61970]: DEBUG nova.network.neutron [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 891.892705] env[61970]: DEBUG nova.compute.manager [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 891.953357] env[61970]: DEBUG nova.network.neutron [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Successfully created port: f836ffa4-f23a-4ddc-8844-4cf747f9e98d {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 892.039803] env[61970]: DEBUG nova.network.neutron [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 892.178764] env[61970]: DEBUG nova.scheduler.client.report [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 892.227507] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]524628ba-6b24-52dc-c9a6-c60d5ff8d351, 'name': SearchDatastore_Task, 'duration_secs': 0.009337} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.228154] env[61970]: DEBUG oslo_concurrency.lockutils [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.228339] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20/8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 892.228605] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6edb7dff-d8cd-498f-8c37-87bd50eeebc2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.234981] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 892.234981] env[61970]: value = "task-1355767" [ 892.234981] env[61970]: _type = "Task" [ 892.234981] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.242405] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355767, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.281313] env[61970]: DEBUG nova.network.neutron [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Updating instance_info_cache with network_info: [{"id": "b83024be-4171-40cd-9d1c-65a2d2ce8b9e", "address": "fa:16:3e:f5:18:c9", "network": {"id": "1f53a96b-8a46-41d5-8bf3-759de1fce443", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-108385347-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a44c05a894394d69b8b4fd36c7b2532a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb83024be-41", "ovs_interfaceid": "b83024be-4171-40cd-9d1c-65a2d2ce8b9e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.303644] env[61970]: DEBUG nova.compute.manager [req-f6a333d4-f1f9-449a-a58e-377ae389dc58 req-50985f21-41c5-44e9-867d-c9321734a8b8 service nova] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Received event network-vif-plugged-b83024be-4171-40cd-9d1c-65a2d2ce8b9e {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.304471] env[61970]: DEBUG oslo_concurrency.lockutils [req-f6a333d4-f1f9-449a-a58e-377ae389dc58 req-50985f21-41c5-44e9-867d-c9321734a8b8 service nova] Acquiring lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.304471] env[61970]: DEBUG oslo_concurrency.lockutils [req-f6a333d4-f1f9-449a-a58e-377ae389dc58 req-50985f21-41c5-44e9-867d-c9321734a8b8 service nova] Lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.304471] env[61970]: DEBUG oslo_concurrency.lockutils [req-f6a333d4-f1f9-449a-a58e-377ae389dc58 req-50985f21-41c5-44e9-867d-c9321734a8b8 service nova] Lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.304471] env[61970]: DEBUG nova.compute.manager [req-f6a333d4-f1f9-449a-a58e-377ae389dc58 req-50985f21-41c5-44e9-867d-c9321734a8b8 service nova] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] No waiting events found dispatching network-vif-plugged-b83024be-4171-40cd-9d1c-65a2d2ce8b9e {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 892.304622] env[61970]: WARNING nova.compute.manager [req-f6a333d4-f1f9-449a-a58e-377ae389dc58 req-50985f21-41c5-44e9-867d-c9321734a8b8 service nova] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Received unexpected event network-vif-plugged-b83024be-4171-40cd-9d1c-65a2d2ce8b9e for instance with vm_state building and task_state spawning. [ 892.304804] env[61970]: DEBUG nova.compute.manager [req-f6a333d4-f1f9-449a-a58e-377ae389dc58 req-50985f21-41c5-44e9-867d-c9321734a8b8 service nova] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Received event network-changed-b83024be-4171-40cd-9d1c-65a2d2ce8b9e {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.304966] env[61970]: DEBUG nova.compute.manager [req-f6a333d4-f1f9-449a-a58e-377ae389dc58 req-50985f21-41c5-44e9-867d-c9321734a8b8 service nova] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Refreshing instance network info cache due to event network-changed-b83024be-4171-40cd-9d1c-65a2d2ce8b9e. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 892.305160] env[61970]: DEBUG oslo_concurrency.lockutils [req-f6a333d4-f1f9-449a-a58e-377ae389dc58 req-50985f21-41c5-44e9-867d-c9321734a8b8 service nova] Acquiring lock "refresh_cache-d0ad9689-bdb7-4d68-b19a-3f805ab4612e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.349029] env[61970]: WARNING nova.network.neutron [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] 8f622fc5-b74f-4582-a6e0-44c90f495750 already exists in list: networks containing: ['8f622fc5-b74f-4582-a6e0-44c90f495750']. ignoring it [ 892.683914] env[61970]: DEBUG oslo_concurrency.lockutils [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.797s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.686276] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 8.716s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.712627] env[61970]: INFO nova.scheduler.client.report [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Deleted allocations for instance 99200b46-7eb7-4ca2-b352-b91b4ad076af [ 892.747452] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355767, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45905} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.747710] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20/8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 892.747924] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 892.748326] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-704ed1e4-c399-4944-a3f4-6380552d6ece {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.755133] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 892.755133] env[61970]: value = "task-1355768" [ 892.755133] env[61970]: _type = "Task" [ 892.755133] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.763287] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355768, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.778929] env[61970]: DEBUG nova.network.neutron [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Updating instance_info_cache with network_info: [{"id": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "address": "fa:16:3e:14:8c:ec", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64a3e684-f5", "ovs_interfaceid": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9", "address": "fa:16:3e:45:a2:e3", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb33aa0c-3f", "ovs_interfaceid": "eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.784150] env[61970]: DEBUG oslo_concurrency.lockutils [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Releasing lock "refresh_cache-d0ad9689-bdb7-4d68-b19a-3f805ab4612e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.784451] env[61970]: DEBUG nova.compute.manager [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Instance network_info: |[{"id": "b83024be-4171-40cd-9d1c-65a2d2ce8b9e", "address": "fa:16:3e:f5:18:c9", "network": {"id": "1f53a96b-8a46-41d5-8bf3-759de1fce443", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-108385347-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a44c05a894394d69b8b4fd36c7b2532a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb83024be-41", "ovs_interfaceid": "b83024be-4171-40cd-9d1c-65a2d2ce8b9e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 892.784741] env[61970]: DEBUG oslo_concurrency.lockutils [req-f6a333d4-f1f9-449a-a58e-377ae389dc58 req-50985f21-41c5-44e9-867d-c9321734a8b8 service nova] Acquired lock "refresh_cache-d0ad9689-bdb7-4d68-b19a-3f805ab4612e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.785275] env[61970]: DEBUG nova.network.neutron [req-f6a333d4-f1f9-449a-a58e-377ae389dc58 req-50985f21-41c5-44e9-867d-c9321734a8b8 service nova] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Refreshing network info cache for port b83024be-4171-40cd-9d1c-65a2d2ce8b9e {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 892.787161] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f5:18:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9bb629cd-6d0f-4bed-965c-bd04a2f3ec49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b83024be-4171-40cd-9d1c-65a2d2ce8b9e', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 892.793482] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Creating folder: Project (a44c05a894394d69b8b4fd36c7b2532a). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 892.794743] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3c38144e-400b-42d2-b050-6a31466457bd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.806042] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Created folder: Project (a44c05a894394d69b8b4fd36c7b2532a) in parent group-v288740. [ 892.806222] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Creating folder: Instances. Parent ref: group-v288817. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 892.806457] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4eed2474-0787-43af-a2a4-c71e297cbb54 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.814820] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Created folder: Instances in parent group-v288817. [ 892.815157] env[61970]: DEBUG oslo.service.loopingcall [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.815247] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 892.815576] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0f981253-6701-479f-a2ab-e3a289205cd4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.833951] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 892.833951] env[61970]: value = "task-1355771" [ 892.833951] env[61970]: _type = "Task" [ 892.833951] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.843972] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355771, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.901437] env[61970]: DEBUG nova.compute.manager [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 892.927299] env[61970]: DEBUG nova.virt.hardware [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 892.927728] env[61970]: DEBUG nova.virt.hardware [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 892.927827] env[61970]: DEBUG nova.virt.hardware [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 892.927986] env[61970]: DEBUG nova.virt.hardware [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 892.928177] env[61970]: DEBUG nova.virt.hardware [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 892.928363] env[61970]: DEBUG nova.virt.hardware [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 892.928581] env[61970]: DEBUG nova.virt.hardware [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 892.928804] env[61970]: DEBUG nova.virt.hardware [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 892.929098] env[61970]: DEBUG nova.virt.hardware [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 892.929188] env[61970]: DEBUG nova.virt.hardware [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 892.929362] env[61970]: DEBUG nova.virt.hardware [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 892.930600] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0530a70-79bf-4667-9c10-9e06605a1a6b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.939933] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac96bd4-aff9-4598-b78d-fc4fa469d281 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.191113] env[61970]: INFO nova.compute.claims [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 893.221274] env[61970]: DEBUG oslo_concurrency.lockutils [None req-325c867f-870c-40da-a8f5-f32324c96ef0 tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "99200b46-7eb7-4ca2-b352-b91b4ad076af" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.178s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.263886] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355768, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.315442} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.264624] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.266458] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4152c0d1-04ca-4b58-9507-fdafd0793ba2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.270051] env[61970]: DEBUG nova.compute.manager [req-6ea0c8ba-1788-4c3b-84e7-5dce2390630c req-d3d0f0ee-4a1f-4cca-b532-0ff6723d912d service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Received event network-changed-eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 893.270248] env[61970]: DEBUG nova.compute.manager [req-6ea0c8ba-1788-4c3b-84e7-5dce2390630c req-d3d0f0ee-4a1f-4cca-b532-0ff6723d912d service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Refreshing instance network info cache due to event network-changed-eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 893.270435] env[61970]: DEBUG oslo_concurrency.lockutils [req-6ea0c8ba-1788-4c3b-84e7-5dce2390630c req-d3d0f0ee-4a1f-4cca-b532-0ff6723d912d service nova] Acquiring lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.283074] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.283835] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.284075] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.292793] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20/8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.293154] env[61970]: DEBUG oslo_concurrency.lockutils [req-6ea0c8ba-1788-4c3b-84e7-5dce2390630c req-d3d0f0ee-4a1f-4cca-b532-0ff6723d912d service nova] Acquired lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.293463] env[61970]: DEBUG nova.network.neutron [req-6ea0c8ba-1788-4c3b-84e7-5dce2390630c req-d3d0f0ee-4a1f-4cca-b532-0ff6723d912d service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Refreshing network info cache for port eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 893.297037] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46846290-0c22-41c0-98d6-1db9e0bee5db {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.298250] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eea16f88-f5c5-4b33-a037-21c7c0bb7296 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.332122] env[61970]: DEBUG nova.virt.hardware [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.332393] env[61970]: DEBUG nova.virt.hardware [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.332549] env[61970]: DEBUG nova.virt.hardware [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.333059] env[61970]: DEBUG nova.virt.hardware [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.333263] env[61970]: DEBUG nova.virt.hardware [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.333420] env[61970]: DEBUG nova.virt.hardware [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.333626] env[61970]: DEBUG nova.virt.hardware [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.333791] env[61970]: DEBUG nova.virt.hardware [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.333947] env[61970]: DEBUG nova.virt.hardware [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.334127] env[61970]: DEBUG nova.virt.hardware [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.334296] env[61970]: DEBUG nova.virt.hardware [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.341524] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Reconfiguring VM to attach interface {{(pid=61970) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 893.343797] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b57a81d-5b28-4d63-9af6-f4f2d4048fcf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.355636] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 893.355636] env[61970]: value = "task-1355772" [ 893.355636] env[61970]: _type = "Task" [ 893.355636] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.365747] env[61970]: DEBUG oslo_vmware.api [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 893.365747] env[61970]: value = "task-1355773" [ 893.365747] env[61970]: _type = "Task" [ 893.365747] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.374201] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355772, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.374393] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355771, 'name': CreateVM_Task, 'duration_secs': 0.394485} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.374849] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 893.375519] env[61970]: DEBUG oslo_concurrency.lockutils [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.375681] env[61970]: DEBUG oslo_concurrency.lockutils [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.376039] env[61970]: DEBUG oslo_concurrency.lockutils [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 893.379336] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb030bb4-7bd5-4907-8870-36597f661c15 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.380947] env[61970]: DEBUG oslo_vmware.api [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355773, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.384015] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 893.384015] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525a1d06-a707-484b-43b4-1e030523b172" [ 893.384015] env[61970]: _type = "Task" [ 893.384015] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.391910] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525a1d06-a707-484b-43b4-1e030523b172, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.686790] env[61970]: DEBUG nova.network.neutron [req-f6a333d4-f1f9-449a-a58e-377ae389dc58 req-50985f21-41c5-44e9-867d-c9321734a8b8 service nova] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Updated VIF entry in instance network info cache for port b83024be-4171-40cd-9d1c-65a2d2ce8b9e. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 893.687313] env[61970]: DEBUG nova.network.neutron [req-f6a333d4-f1f9-449a-a58e-377ae389dc58 req-50985f21-41c5-44e9-867d-c9321734a8b8 service nova] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Updating instance_info_cache with network_info: [{"id": "b83024be-4171-40cd-9d1c-65a2d2ce8b9e", "address": "fa:16:3e:f5:18:c9", "network": {"id": "1f53a96b-8a46-41d5-8bf3-759de1fce443", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-108385347-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a44c05a894394d69b8b4fd36c7b2532a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb83024be-41", "ovs_interfaceid": "b83024be-4171-40cd-9d1c-65a2d2ce8b9e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.697904] env[61970]: INFO nova.compute.resource_tracker [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Updating resource usage from migration 1759ca51-3a57-4c63-81ae-f85ae6bc5d5f [ 893.875506] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355772, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.880811] env[61970]: DEBUG oslo_vmware.api [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355773, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.898540] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525a1d06-a707-484b-43b4-1e030523b172, 'name': SearchDatastore_Task, 'duration_secs': 0.022678} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.898838] env[61970]: DEBUG oslo_concurrency.lockutils [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.899080] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 893.899316] env[61970]: DEBUG oslo_concurrency.lockutils [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.899461] env[61970]: DEBUG oslo_concurrency.lockutils [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.899637] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 893.899892] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cd95d93f-654a-4a60-bce4-3ee1d7f86482 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.906111] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d30352d-361a-47b8-a579-69a2eaf9d721 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.910244] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 893.910420] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 893.912531] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bb90ed9-1c51-47c0-a321-d624223c7086 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.916539] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65288165-5fbb-48be-8d4a-bf289eb4082d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.920578] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 893.920578] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52cf0686-31f1-ac14-435b-16ea965675d2" [ 893.920578] env[61970]: _type = "Task" [ 893.920578] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.954414] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Acquiring lock "b385407b-1bdd-4c53-907c-cb4c8ce16cc7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.954671] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Lock "b385407b-1bdd-4c53-907c-cb4c8ce16cc7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.954869] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Acquiring lock "b385407b-1bdd-4c53-907c-cb4c8ce16cc7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.955072] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Lock "b385407b-1bdd-4c53-907c-cb4c8ce16cc7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.955245] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Lock "b385407b-1bdd-4c53-907c-cb4c8ce16cc7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.961055] env[61970]: INFO nova.compute.manager [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Terminating instance [ 893.963690] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a5be6df-9979-4973-a23e-3148071a3008 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.966938] env[61970]: DEBUG nova.compute.manager [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 893.967213] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 893.971906] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f4b70b-d878-4a17-9e68-4f08553fe877 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.978119] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52cf0686-31f1-ac14-435b-16ea965675d2, 'name': SearchDatastore_Task, 'duration_secs': 0.007914} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.978119] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9f310a2-d563-4a2c-bdff-8077ec0f43ea {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.982851] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 893.984580] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd1b3d9f-9e23-4532-a569-7261461c39cb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.987928] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bac8a0f2-ef6d-4a5f-824c-d74d3e12dc16 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.989844] env[61970]: DEBUG nova.network.neutron [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Successfully updated port: f836ffa4-f23a-4ddc-8844-4cf747f9e98d {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 893.992181] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 893.992181] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d643b0-7687-07bc-3af9-764e4f5a8ab3" [ 893.992181] env[61970]: _type = "Task" [ 893.992181] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.003929] env[61970]: DEBUG nova.compute.provider_tree [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.010590] env[61970]: DEBUG oslo_vmware.api [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Waiting for the task: (returnval){ [ 894.010590] env[61970]: value = "task-1355774" [ 894.010590] env[61970]: _type = "Task" [ 894.010590] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.017993] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d643b0-7687-07bc-3af9-764e4f5a8ab3, 'name': SearchDatastore_Task, 'duration_secs': 0.007623} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.018634] env[61970]: DEBUG oslo_concurrency.lockutils [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.018913] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] d0ad9689-bdb7-4d68-b19a-3f805ab4612e/d0ad9689-bdb7-4d68-b19a-3f805ab4612e.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 894.019420] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8c088161-eb9a-45a7-aef5-5fd2f2de4278 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.024858] env[61970]: DEBUG oslo_vmware.api [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': task-1355774, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.029643] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 894.029643] env[61970]: value = "task-1355775" [ 894.029643] env[61970]: _type = "Task" [ 894.029643] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.038225] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355775, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.159620] env[61970]: DEBUG nova.network.neutron [req-6ea0c8ba-1788-4c3b-84e7-5dce2390630c req-d3d0f0ee-4a1f-4cca-b532-0ff6723d912d service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Updated VIF entry in instance network info cache for port eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 894.160191] env[61970]: DEBUG nova.network.neutron [req-6ea0c8ba-1788-4c3b-84e7-5dce2390630c req-d3d0f0ee-4a1f-4cca-b532-0ff6723d912d service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Updating instance_info_cache with network_info: [{"id": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "address": "fa:16:3e:14:8c:ec", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64a3e684-f5", "ovs_interfaceid": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9", "address": "fa:16:3e:45:a2:e3", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb33aa0c-3f", "ovs_interfaceid": "eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.189773] env[61970]: DEBUG oslo_concurrency.lockutils [req-f6a333d4-f1f9-449a-a58e-377ae389dc58 req-50985f21-41c5-44e9-867d-c9321734a8b8 service nova] Releasing lock "refresh_cache-d0ad9689-bdb7-4d68-b19a-3f805ab4612e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.375714] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355772, 'name': ReconfigVM_Task, 'duration_secs': 0.79408} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.376543] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20/8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.377273] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d7d8e37e-ade1-4d0f-9057-cdae80132998 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.382308] env[61970]: DEBUG oslo_vmware.api [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355773, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.387318] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 894.387318] env[61970]: value = "task-1355776" [ 894.387318] env[61970]: _type = "Task" [ 894.387318] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.398757] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355776, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.494355] env[61970]: DEBUG oslo_concurrency.lockutils [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "refresh_cache-0c50937e-4646-47ec-ac0c-0281c9424d86" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.494577] env[61970]: DEBUG oslo_concurrency.lockutils [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "refresh_cache-0c50937e-4646-47ec-ac0c-0281c9424d86" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.494818] env[61970]: DEBUG nova.network.neutron [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 894.511127] env[61970]: DEBUG nova.scheduler.client.report [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 894.525393] env[61970]: DEBUG oslo_vmware.api [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': task-1355774, 'name': PowerOffVM_Task, 'duration_secs': 0.376617} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.525659] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 894.525873] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 894.526224] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0bbb64c3-6733-49ec-b8dd-764e97100ce8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.539812] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355775, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.620170] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 894.620394] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 894.620572] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Deleting the datastore file [datastore1] b385407b-1bdd-4c53-907c-cb4c8ce16cc7 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 894.620843] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-334a85b5-5263-4d25-9888-2fbaa0bac4cf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.627775] env[61970]: DEBUG oslo_vmware.api [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Waiting for the task: (returnval){ [ 894.627775] env[61970]: value = "task-1355778" [ 894.627775] env[61970]: _type = "Task" [ 894.627775] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.635796] env[61970]: DEBUG oslo_vmware.api [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': task-1355778, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.662867] env[61970]: DEBUG oslo_concurrency.lockutils [req-6ea0c8ba-1788-4c3b-84e7-5dce2390630c req-d3d0f0ee-4a1f-4cca-b532-0ff6723d912d service nova] Releasing lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.877937] env[61970]: DEBUG oslo_vmware.api [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355773, 'name': ReconfigVM_Task, 'duration_secs': 1.031139} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.878538] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.878755] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Reconfigured VM to attach interface {{(pid=61970) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 894.896396] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355776, 'name': Rename_Task, 'duration_secs': 0.290936} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.896653] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 894.897565] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5bddc9d2-f1f8-4b59-acdf-357504596ac9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.904401] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 894.904401] env[61970]: value = "task-1355779" [ 894.904401] env[61970]: _type = "Task" [ 894.904401] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.911872] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355779, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.016443] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.330s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.016761] env[61970]: INFO nova.compute.manager [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Migrating [ 895.017203] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.017431] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.019100] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.419s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.019410] env[61970]: DEBUG nova.objects.instance [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lazy-loading 'resources' on Instance uuid 2d280f75-ed84-4db3-9067-8ccc719b0d2e {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.026919] env[61970]: DEBUG nova.network.neutron [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 895.041037] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355775, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522914} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.043261] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] d0ad9689-bdb7-4d68-b19a-3f805ab4612e/d0ad9689-bdb7-4d68-b19a-3f805ab4612e.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 895.043482] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 895.043744] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7fff864c-d39f-4ce1-baa7-48d42eb66e9b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.050501] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 895.050501] env[61970]: value = "task-1355780" [ 895.050501] env[61970]: _type = "Task" [ 895.050501] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.058273] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355780, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.137480] env[61970]: DEBUG oslo_vmware.api [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Task: {'id': task-1355778, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.246902} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.137714] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.137896] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 895.138085] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 895.138361] env[61970]: INFO nova.compute.manager [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Took 1.17 seconds to destroy the instance on the hypervisor. [ 895.138495] env[61970]: DEBUG oslo.service.loopingcall [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.138680] env[61970]: DEBUG nova.compute.manager [-] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 895.138776] env[61970]: DEBUG nova.network.neutron [-] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 895.259440] env[61970]: DEBUG nova.network.neutron [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Updating instance_info_cache with network_info: [{"id": "f836ffa4-f23a-4ddc-8844-4cf747f9e98d", "address": "fa:16:3e:38:a2:66", "network": {"id": "62edd125-b77c-4418-a465-1dcb3a0733c1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-371307366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6a4997482a4834a30c79c4cbab2cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf836ffa4-f2", "ovs_interfaceid": "f836ffa4-f23a-4ddc-8844-4cf747f9e98d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.387453] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f57651ce-6a32-4a9b-854c-e6b5013f19da tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "interface-b1cc1cc2-15d6-459d-9529-e592ddb225ac-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.679s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.415323] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355779, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.467958] env[61970]: DEBUG nova.compute.manager [req-00f29179-5b3b-4827-bcbc-ccdd09073681 req-e81439a0-9ad3-4e77-b020-443c3aa4b9a6 service nova] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Received event network-vif-plugged-f836ffa4-f23a-4ddc-8844-4cf747f9e98d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 895.468121] env[61970]: DEBUG oslo_concurrency.lockutils [req-00f29179-5b3b-4827-bcbc-ccdd09073681 req-e81439a0-9ad3-4e77-b020-443c3aa4b9a6 service nova] Acquiring lock "0c50937e-4646-47ec-ac0c-0281c9424d86-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.468328] env[61970]: DEBUG oslo_concurrency.lockutils [req-00f29179-5b3b-4827-bcbc-ccdd09073681 req-e81439a0-9ad3-4e77-b020-443c3aa4b9a6 service nova] Lock "0c50937e-4646-47ec-ac0c-0281c9424d86-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.468492] env[61970]: DEBUG oslo_concurrency.lockutils [req-00f29179-5b3b-4827-bcbc-ccdd09073681 req-e81439a0-9ad3-4e77-b020-443c3aa4b9a6 service nova] Lock "0c50937e-4646-47ec-ac0c-0281c9424d86-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.468665] env[61970]: DEBUG nova.compute.manager [req-00f29179-5b3b-4827-bcbc-ccdd09073681 req-e81439a0-9ad3-4e77-b020-443c3aa4b9a6 service nova] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] No waiting events found dispatching network-vif-plugged-f836ffa4-f23a-4ddc-8844-4cf747f9e98d {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 895.469186] env[61970]: WARNING nova.compute.manager [req-00f29179-5b3b-4827-bcbc-ccdd09073681 req-e81439a0-9ad3-4e77-b020-443c3aa4b9a6 service nova] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Received unexpected event network-vif-plugged-f836ffa4-f23a-4ddc-8844-4cf747f9e98d for instance with vm_state building and task_state spawning. [ 895.469186] env[61970]: DEBUG nova.compute.manager [req-00f29179-5b3b-4827-bcbc-ccdd09073681 req-e81439a0-9ad3-4e77-b020-443c3aa4b9a6 service nova] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Received event network-changed-f836ffa4-f23a-4ddc-8844-4cf747f9e98d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 895.469186] env[61970]: DEBUG nova.compute.manager [req-00f29179-5b3b-4827-bcbc-ccdd09073681 req-e81439a0-9ad3-4e77-b020-443c3aa4b9a6 service nova] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Refreshing instance network info cache due to event network-changed-f836ffa4-f23a-4ddc-8844-4cf747f9e98d. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 895.469422] env[61970]: DEBUG oslo_concurrency.lockutils [req-00f29179-5b3b-4827-bcbc-ccdd09073681 req-e81439a0-9ad3-4e77-b020-443c3aa4b9a6 service nova] Acquiring lock "refresh_cache-0c50937e-4646-47ec-ac0c-0281c9424d86" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.521878] env[61970]: INFO nova.compute.rpcapi [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 895.522485] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.570756] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355780, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072811} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.571447] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 895.572165] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6536ac9-ad87-4603-bbc7-0965bbf6025f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.596849] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] d0ad9689-bdb7-4d68-b19a-3f805ab4612e/d0ad9689-bdb7-4d68-b19a-3f805ab4612e.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 895.597456] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56c0cb9f-c3ea-45ef-bcc7-f2dc2c0f53bf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.620908] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 895.620908] env[61970]: value = "task-1355781" [ 895.620908] env[61970]: _type = "Task" [ 895.620908] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.633223] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355781, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.759704] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a73fe6f-15f2-4969-b152-2d04bd6a35ea {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.762864] env[61970]: DEBUG oslo_concurrency.lockutils [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "refresh_cache-0c50937e-4646-47ec-ac0c-0281c9424d86" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.763172] env[61970]: DEBUG nova.compute.manager [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Instance network_info: |[{"id": "f836ffa4-f23a-4ddc-8844-4cf747f9e98d", "address": "fa:16:3e:38:a2:66", "network": {"id": "62edd125-b77c-4418-a465-1dcb3a0733c1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-371307366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6a4997482a4834a30c79c4cbab2cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf836ffa4-f2", "ovs_interfaceid": "f836ffa4-f23a-4ddc-8844-4cf747f9e98d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 895.763453] env[61970]: DEBUG oslo_concurrency.lockutils [req-00f29179-5b3b-4827-bcbc-ccdd09073681 req-e81439a0-9ad3-4e77-b020-443c3aa4b9a6 service nova] Acquired lock "refresh_cache-0c50937e-4646-47ec-ac0c-0281c9424d86" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.763626] env[61970]: DEBUG nova.network.neutron [req-00f29179-5b3b-4827-bcbc-ccdd09073681 req-e81439a0-9ad3-4e77-b020-443c3aa4b9a6 service nova] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Refreshing network info cache for port f836ffa4-f23a-4ddc-8844-4cf747f9e98d {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 895.764897] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:a2:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f836ffa4-f23a-4ddc-8844-4cf747f9e98d', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 895.772319] env[61970]: DEBUG oslo.service.loopingcall [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.776121] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 895.777936] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b9c91889-ca2c-4dea-91e5-7ff0975614b5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.794058] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-270482d7-d5f4-42ee-97df-8b5ea3c64291 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.832471] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d8ba9d7-f5ea-4465-a466-b6fad174cb2e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.835640] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 895.835640] env[61970]: value = "task-1355782" [ 895.835640] env[61970]: _type = "Task" [ 895.835640] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.848054] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992a5508-c5b7-4617-b4e6-3882b93416ab {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.855245] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355782, 'name': CreateVM_Task} progress is 15%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.866029] env[61970]: DEBUG nova.compute.provider_tree [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.915909] env[61970]: DEBUG oslo_vmware.api [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355779, 'name': PowerOnVM_Task, 'duration_secs': 0.776809} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.916211] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 895.916708] env[61970]: INFO nova.compute.manager [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Took 8.59 seconds to spawn the instance on the hypervisor. [ 895.916708] env[61970]: DEBUG nova.compute.manager [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.917870] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d77468-170d-43ab-9eee-cc274af83585 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.043137] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "refresh_cache-75762ba3-f129-424e-a7cb-962785dfab2a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.043448] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "refresh_cache-75762ba3-f129-424e-a7cb-962785dfab2a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.043540] env[61970]: DEBUG nova.network.neutron [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 896.081131] env[61970]: DEBUG nova.network.neutron [req-00f29179-5b3b-4827-bcbc-ccdd09073681 req-e81439a0-9ad3-4e77-b020-443c3aa4b9a6 service nova] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Updated VIF entry in instance network info cache for port f836ffa4-f23a-4ddc-8844-4cf747f9e98d. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 896.081511] env[61970]: DEBUG nova.network.neutron [req-00f29179-5b3b-4827-bcbc-ccdd09073681 req-e81439a0-9ad3-4e77-b020-443c3aa4b9a6 service nova] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Updating instance_info_cache with network_info: [{"id": "f836ffa4-f23a-4ddc-8844-4cf747f9e98d", "address": "fa:16:3e:38:a2:66", "network": {"id": "62edd125-b77c-4418-a465-1dcb3a0733c1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-371307366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6a4997482a4834a30c79c4cbab2cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf836ffa4-f2", "ovs_interfaceid": "f836ffa4-f23a-4ddc-8844-4cf747f9e98d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.131235] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355781, 'name': ReconfigVM_Task, 'duration_secs': 0.301952} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.131506] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Reconfigured VM instance instance-0000004e to attach disk [datastore2] d0ad9689-bdb7-4d68-b19a-3f805ab4612e/d0ad9689-bdb7-4d68-b19a-3f805ab4612e.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 896.132222] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-428f102b-bd88-4d49-af31-0a0827623072 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.138316] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 896.138316] env[61970]: value = "task-1355783" [ 896.138316] env[61970]: _type = "Task" [ 896.138316] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.147256] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355783, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.339579] env[61970]: DEBUG nova.network.neutron [-] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.350466] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355782, 'name': CreateVM_Task, 'duration_secs': 0.406861} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.350466] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 896.351204] env[61970]: DEBUG oslo_concurrency.lockutils [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.351372] env[61970]: DEBUG oslo_concurrency.lockutils [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.352341] env[61970]: DEBUG oslo_concurrency.lockutils [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 896.353284] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e1bdcb4-40be-4928-9d48-0a2293251af4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.358287] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 896.358287] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529ab9c9-dacd-6ea4-bced-9fcf39a8492c" [ 896.358287] env[61970]: _type = "Task" [ 896.358287] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.367430] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529ab9c9-dacd-6ea4-bced-9fcf39a8492c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.369458] env[61970]: DEBUG nova.scheduler.client.report [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 896.438810] env[61970]: INFO nova.compute.manager [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Took 16.78 seconds to build instance. [ 896.583888] env[61970]: DEBUG oslo_concurrency.lockutils [req-00f29179-5b3b-4827-bcbc-ccdd09073681 req-e81439a0-9ad3-4e77-b020-443c3aa4b9a6 service nova] Releasing lock "refresh_cache-0c50937e-4646-47ec-ac0c-0281c9424d86" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.647831] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355783, 'name': Rename_Task, 'duration_secs': 0.142231} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.648640] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 896.648640] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e476ba2e-bd5b-4964-8780-9b2fac365336 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.655565] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 896.655565] env[61970]: value = "task-1355784" [ 896.655565] env[61970]: _type = "Task" [ 896.655565] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.663397] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355784, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.845127] env[61970]: INFO nova.compute.manager [-] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Took 1.71 seconds to deallocate network for instance. [ 896.872361] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529ab9c9-dacd-6ea4-bced-9fcf39a8492c, 'name': SearchDatastore_Task, 'duration_secs': 0.012954} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.875806] env[61970]: DEBUG oslo_concurrency.lockutils [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.876072] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 896.876343] env[61970]: DEBUG oslo_concurrency.lockutils [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.876496] env[61970]: DEBUG oslo_concurrency.lockutils [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.876681] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 896.877398] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.858s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.879667] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-17654a81-1a00-4305-8a41-d7eeae91c217 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.882770] env[61970]: DEBUG oslo_concurrency.lockutils [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.775s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.882770] env[61970]: DEBUG nova.objects.instance [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lazy-loading 'resources' on Instance uuid e2185ed0-2bc1-4718-b47d-57150b5e60ba {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.890386] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 896.890600] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 896.892271] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-294d0d04-11f7-469f-bac6-663d229b605d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.898292] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 896.898292] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52cc0d77-2b39-137a-b5d3-f012fe840c12" [ 896.898292] env[61970]: _type = "Task" [ 896.898292] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.906114] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52cc0d77-2b39-137a-b5d3-f012fe840c12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.907083] env[61970]: INFO nova.scheduler.client.report [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Deleted allocations for instance 2d280f75-ed84-4db3-9067-8ccc719b0d2e [ 896.942015] env[61970]: DEBUG oslo_concurrency.lockutils [None req-652eb234-b7f2-4782-8d5f-ddd03a435878 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.297s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.079497] env[61970]: DEBUG oslo_concurrency.lockutils [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.079766] env[61970]: DEBUG oslo_concurrency.lockutils [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.079986] env[61970]: DEBUG oslo_concurrency.lockutils [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.080202] env[61970]: DEBUG oslo_concurrency.lockutils [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.080396] env[61970]: DEBUG oslo_concurrency.lockutils [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.083097] env[61970]: INFO nova.compute.manager [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Terminating instance [ 897.085085] env[61970]: DEBUG nova.compute.manager [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 897.085820] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 897.087181] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b306f8c4-19ad-4da4-b201-816299b6c89f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.096525] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 897.096779] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2061278d-ea95-4903-b7ca-d3250de1de73 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.103204] env[61970]: DEBUG nova.network.neutron [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Updating instance_info_cache with network_info: [{"id": "32711b6a-a9cc-472d-a634-e3b60711bffe", "address": "fa:16:3e:68:3d:bc", "network": {"id": "cba185ff-bfa3-4547-b7b0-638c3e79a24c", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-864852491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5c64cb9232fb413cbd7627dcf077e9ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32711b6a-a9", "ovs_interfaceid": "32711b6a-a9cc-472d-a634-e3b60711bffe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.109887] env[61970]: DEBUG oslo_vmware.api [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 897.109887] env[61970]: value = "task-1355785" [ 897.109887] env[61970]: _type = "Task" [ 897.109887] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.121256] env[61970]: DEBUG oslo_vmware.api [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355785, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.166047] env[61970]: DEBUG oslo_vmware.api [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355784, 'name': PowerOnVM_Task, 'duration_secs': 0.449492} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.166325] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 897.166545] env[61970]: INFO nova.compute.manager [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Took 6.56 seconds to spawn the instance on the hypervisor. [ 897.166745] env[61970]: DEBUG nova.compute.manager [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.167738] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7cad9d0-4fcc-492f-b22b-fed66381e6a5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.355118] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.421768] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52cc0d77-2b39-137a-b5d3-f012fe840c12, 'name': SearchDatastore_Task, 'duration_secs': 0.008877} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.422945] env[61970]: DEBUG oslo_concurrency.lockutils [None req-99a845d3-6a5f-46b8-bc18-4ac461ee1cde tempest-ImagesTestJSON-37838668 tempest-ImagesTestJSON-37838668-project-member] Lock "2d280f75-ed84-4db3-9067-8ccc719b0d2e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.866s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.425499] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb4ecc77-4d59-4c6c-8021-36a87c917672 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.435702] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 897.435702] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5268bb9d-4b28-8ff0-cdda-7e38421df498" [ 897.435702] env[61970]: _type = "Task" [ 897.435702] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.444843] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5268bb9d-4b28-8ff0-cdda-7e38421df498, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.564178] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f553a16-ecf2-4505-863b-3cf1739f3399 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.571742] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029562ea-f34d-4939-8bc6-10996fee3f21 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.601801] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26f3b3c5-cacc-4bdb-af4f-ff79b13ae457 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.605314] env[61970]: DEBUG nova.compute.manager [req-12609b2e-7721-4605-aec3-43ed19fb1e8c req-94e12927-b8d0-4db0-b6d4-352a0f29fbd5 service nova] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Received event network-vif-deleted-6875c46d-b7c5-4912-a1f0-932b3ab900d8 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 897.610083] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "refresh_cache-75762ba3-f129-424e-a7cb-962785dfab2a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.612489] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be987ba9-0524-4c3a-bcdd-e61760b5c8e4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.629015] env[61970]: DEBUG nova.compute.provider_tree [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.633545] env[61970]: DEBUG oslo_vmware.api [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355785, 'name': PowerOffVM_Task, 'duration_secs': 0.214672} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.633951] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 897.634152] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 897.634402] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-98502403-69e9-4aac-b26c-a119ac8a3147 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.686897] env[61970]: INFO nova.compute.manager [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Took 17.38 seconds to build instance. [ 897.695052] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "interface-b1cc1cc2-15d6-459d-9529-e592ddb225ac-b47e01b1-0321-4bbd-8ce3-d19568926d99" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.695052] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "interface-b1cc1cc2-15d6-459d-9529-e592ddb225ac-b47e01b1-0321-4bbd-8ce3-d19568926d99" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.695052] env[61970]: DEBUG nova.objects.instance [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lazy-loading 'flavor' on Instance uuid b1cc1cc2-15d6-459d-9529-e592ddb225ac {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.723950] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 897.724641] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 897.724641] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleting the datastore file [datastore2] 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 897.724721] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-48dc9c45-8bd8-406a-be7a-9fef643cca45 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.731167] env[61970]: DEBUG oslo_vmware.api [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 897.731167] env[61970]: value = "task-1355787" [ 897.731167] env[61970]: _type = "Task" [ 897.731167] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.738612] env[61970]: DEBUG oslo_vmware.api [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355787, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.919676] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Acquiring lock "6838c6a4-8241-4b6f-ad58-3acfd7afdbf4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.919946] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Lock "6838c6a4-8241-4b6f-ad58-3acfd7afdbf4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.920180] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Acquiring lock "6838c6a4-8241-4b6f-ad58-3acfd7afdbf4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.920366] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Lock "6838c6a4-8241-4b6f-ad58-3acfd7afdbf4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.920590] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Lock "6838c6a4-8241-4b6f-ad58-3acfd7afdbf4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.922628] env[61970]: INFO nova.compute.manager [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Terminating instance [ 897.924435] env[61970]: DEBUG nova.compute.manager [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 897.924713] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 897.924864] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c402b44-c4f2-4f50-804e-2307562f1cbf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.931561] env[61970]: DEBUG oslo_vmware.api [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Waiting for the task: (returnval){ [ 897.931561] env[61970]: value = "task-1355788" [ 897.931561] env[61970]: _type = "Task" [ 897.931561] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.941921] env[61970]: DEBUG oslo_vmware.api [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355788, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.947777] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5268bb9d-4b28-8ff0-cdda-7e38421df498, 'name': SearchDatastore_Task, 'duration_secs': 0.018591} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.947777] env[61970]: DEBUG oslo_concurrency.lockutils [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.947777] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 0c50937e-4646-47ec-ac0c-0281c9424d86/0c50937e-4646-47ec-ac0c-0281c9424d86.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 897.947777] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-89e7289b-8b9f-4782-9785-2f489900e4bd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.953469] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 897.953469] env[61970]: value = "task-1355789" [ 897.953469] env[61970]: _type = "Task" [ 897.953469] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.968775] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355789, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.135346] env[61970]: DEBUG nova.scheduler.client.report [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 898.189434] env[61970]: DEBUG oslo_concurrency.lockutils [None req-54aedbdb-4b31-4618-b519-5c17066abeb1 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.893s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.245041] env[61970]: DEBUG oslo_vmware.api [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355787, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14681} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.248379] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 898.248379] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 898.248379] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 898.248379] env[61970]: INFO nova.compute.manager [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Took 1.16 seconds to destroy the instance on the hypervisor. [ 898.248379] env[61970]: DEBUG oslo.service.loopingcall [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 898.248379] env[61970]: DEBUG nova.compute.manager [-] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 898.248379] env[61970]: DEBUG nova.network.neutron [-] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 898.442659] env[61970]: DEBUG oslo_vmware.api [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355788, 'name': PowerOffVM_Task, 'duration_secs': 0.226033} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.442994] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 898.443251] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Volume detach. Driver type: vmdk {{(pid=61970) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 898.443485] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288764', 'volume_id': '4020284d-a273-459a-bb8c-84753ab98b1e', 'name': 'volume-4020284d-a273-459a-bb8c-84753ab98b1e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6838c6a4-8241-4b6f-ad58-3acfd7afdbf4', 'attached_at': '', 'detached_at': '', 'volume_id': '4020284d-a273-459a-bb8c-84753ab98b1e', 'serial': '4020284d-a273-459a-bb8c-84753ab98b1e'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 898.444467] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b73afa44-74b4-49c3-81b0-e4280daa2e02 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.467268] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f82f1bd-2322-4c27-b50c-f9c6c9f3db88 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.475992] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355789, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.482185] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a0abcb1-06ee-48d2-bb90-2a415e63aa59 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.500752] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e7e761-e288-4304-abfa-fe262d4fc56d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.506356] env[61970]: DEBUG nova.objects.instance [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lazy-loading 'pci_requests' on Instance uuid b1cc1cc2-15d6-459d-9529-e592ddb225ac {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.520665] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] The volume has not been displaced from its original location: [datastore1] volume-4020284d-a273-459a-bb8c-84753ab98b1e/volume-4020284d-a273-459a-bb8c-84753ab98b1e.vmdk. No consolidation needed. {{(pid=61970) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 898.526171] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Reconfiguring VM instance instance-00000043 to detach disk 2000 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 898.527810] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68c6c83f-aee7-4fa8-a045-d21e1e22168f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.546851] env[61970]: DEBUG oslo_vmware.api [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Waiting for the task: (returnval){ [ 898.546851] env[61970]: value = "task-1355790" [ 898.546851] env[61970]: _type = "Task" [ 898.546851] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.555486] env[61970]: DEBUG oslo_vmware.api [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355790, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.639900] env[61970]: DEBUG oslo_concurrency.lockutils [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.758s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.643126] env[61970]: DEBUG oslo_concurrency.lockutils [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.504s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.643126] env[61970]: DEBUG nova.objects.instance [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lazy-loading 'resources' on Instance uuid 5d4d97e1-5812-4884-a1ea-6eb0cee591c0 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.663062] env[61970]: INFO nova.scheduler.client.report [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Deleted allocations for instance e2185ed0-2bc1-4718-b47d-57150b5e60ba [ 898.936658] env[61970]: DEBUG nova.compute.manager [req-1a49adab-df88-4876-93ac-62ec66f85dbf req-91ba6432-2836-4403-8bba-dabeafd0f91a service nova] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Received event network-changed-b83024be-4171-40cd-9d1c-65a2d2ce8b9e {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.936866] env[61970]: DEBUG nova.compute.manager [req-1a49adab-df88-4876-93ac-62ec66f85dbf req-91ba6432-2836-4403-8bba-dabeafd0f91a service nova] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Refreshing instance network info cache due to event network-changed-b83024be-4171-40cd-9d1c-65a2d2ce8b9e. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 898.937092] env[61970]: DEBUG oslo_concurrency.lockutils [req-1a49adab-df88-4876-93ac-62ec66f85dbf req-91ba6432-2836-4403-8bba-dabeafd0f91a service nova] Acquiring lock "refresh_cache-d0ad9689-bdb7-4d68-b19a-3f805ab4612e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.937239] env[61970]: DEBUG oslo_concurrency.lockutils [req-1a49adab-df88-4876-93ac-62ec66f85dbf req-91ba6432-2836-4403-8bba-dabeafd0f91a service nova] Acquired lock "refresh_cache-d0ad9689-bdb7-4d68-b19a-3f805ab4612e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.937416] env[61970]: DEBUG nova.network.neutron [req-1a49adab-df88-4876-93ac-62ec66f85dbf req-91ba6432-2836-4403-8bba-dabeafd0f91a service nova] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Refreshing network info cache for port b83024be-4171-40cd-9d1c-65a2d2ce8b9e {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 898.981643] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355789, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.589038} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.984012] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 0c50937e-4646-47ec-ac0c-0281c9424d86/0c50937e-4646-47ec-ac0c-0281c9424d86.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 898.984012] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 898.984012] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8d483e02-6202-4402-b174-f156ff68f52f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.991053] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 898.991053] env[61970]: value = "task-1355791" [ 898.991053] env[61970]: _type = "Task" [ 898.991053] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.002855] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355791, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.027613] env[61970]: DEBUG nova.objects.base [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 899.027828] env[61970]: DEBUG nova.network.neutron [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 899.056306] env[61970]: DEBUG oslo_vmware.api [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355790, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.130232] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b2b754-702d-4e81-84d0-c180805aa497 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.153169] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Updating instance '75762ba3-f129-424e-a7cb-962785dfab2a' progress to 0 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 899.174610] env[61970]: DEBUG oslo_concurrency.lockutils [None req-cb2a774a-5099-41c8-9bd0-d1656d773d2e tempest-MultipleCreateTestJSON-1314765417 tempest-MultipleCreateTestJSON-1314765417-project-member] Lock "e2185ed0-2bc1-4718-b47d-57150b5e60ba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.054s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.197528] env[61970]: DEBUG nova.policy [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2af915dc43c84940a937dfb8a04ea0b6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17ca05a1e6664430a02de563d98c1148', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 899.340348] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48077296-7524-4426-aa8b-74127d1e9611 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.348706] env[61970]: DEBUG nova.network.neutron [-] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.351359] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e00338b-d975-4fc1-b5ce-bc4c6f35081e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.383431] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9ceae7c-eb33-47b3-965e-fe8fb36242e7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.392836] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2f5fd0-dc51-4a6a-a56c-406c38cbb21e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.408069] env[61970]: DEBUG nova.compute.provider_tree [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 899.500449] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355791, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.557720] env[61970]: DEBUG oslo_vmware.api [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355790, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.659206] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 899.659589] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52a02582-34f2-401d-ab18-f820c9903c18 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.666422] env[61970]: DEBUG oslo_vmware.api [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 899.666422] env[61970]: value = "task-1355792" [ 899.666422] env[61970]: _type = "Task" [ 899.666422] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.677238] env[61970]: DEBUG oslo_vmware.api [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355792, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.707707] env[61970]: DEBUG nova.network.neutron [req-1a49adab-df88-4876-93ac-62ec66f85dbf req-91ba6432-2836-4403-8bba-dabeafd0f91a service nova] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Updated VIF entry in instance network info cache for port b83024be-4171-40cd-9d1c-65a2d2ce8b9e. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 899.708181] env[61970]: DEBUG nova.network.neutron [req-1a49adab-df88-4876-93ac-62ec66f85dbf req-91ba6432-2836-4403-8bba-dabeafd0f91a service nova] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Updating instance_info_cache with network_info: [{"id": "b83024be-4171-40cd-9d1c-65a2d2ce8b9e", "address": "fa:16:3e:f5:18:c9", "network": {"id": "1f53a96b-8a46-41d5-8bf3-759de1fce443", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-108385347-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a44c05a894394d69b8b4fd36c7b2532a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb83024be-41", "ovs_interfaceid": "b83024be-4171-40cd-9d1c-65a2d2ce8b9e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.794432] env[61970]: DEBUG nova.compute.manager [req-d6eec787-66a3-446c-a34e-fde71759aa66 req-08f8b438-d959-4cbd-9520-163637e0eed8 service nova] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Received event network-vif-deleted-7d01e768-116a-4e5c-b85a-8a9920781421 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.856118] env[61970]: INFO nova.compute.manager [-] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Took 1.61 seconds to deallocate network for instance. [ 899.911176] env[61970]: DEBUG nova.scheduler.client.report [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 900.002433] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355791, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.655081} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.002672] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 900.003454] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d98cb0-4b3a-46fc-b07f-fb8435cc943a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.028352] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 0c50937e-4646-47ec-ac0c-0281c9424d86/0c50937e-4646-47ec-ac0c-0281c9424d86.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 900.028703] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c281e04-d9af-4858-bb5b-f8694c5ca8d9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.048959] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 900.048959] env[61970]: value = "task-1355793" [ 900.048959] env[61970]: _type = "Task" [ 900.048959] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.060687] env[61970]: DEBUG oslo_vmware.api [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355790, 'name': ReconfigVM_Task, 'duration_secs': 1.226452} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.063815] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Reconfigured VM instance instance-00000043 to detach disk 2000 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 900.068508] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355793, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.069109] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ecddbfd1-4326-4cdd-ab0a-e849cd4dfbd5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.083911] env[61970]: DEBUG oslo_vmware.api [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Waiting for the task: (returnval){ [ 900.083911] env[61970]: value = "task-1355794" [ 900.083911] env[61970]: _type = "Task" [ 900.083911] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.092387] env[61970]: DEBUG oslo_vmware.api [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355794, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.179612] env[61970]: DEBUG oslo_vmware.api [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355792, 'name': PowerOffVM_Task, 'duration_secs': 0.185478} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.180578] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 900.180578] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Updating instance '75762ba3-f129-424e-a7cb-962785dfab2a' progress to 17 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 900.211662] env[61970]: DEBUG oslo_concurrency.lockutils [req-1a49adab-df88-4876-93ac-62ec66f85dbf req-91ba6432-2836-4403-8bba-dabeafd0f91a service nova] Releasing lock "refresh_cache-d0ad9689-bdb7-4d68-b19a-3f805ab4612e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.375453] env[61970]: DEBUG oslo_concurrency.lockutils [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.417447] env[61970]: DEBUG oslo_concurrency.lockutils [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.775s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.419876] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.829s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.421509] env[61970]: INFO nova.compute.claims [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 900.448337] env[61970]: INFO nova.scheduler.client.report [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Deleted allocations for instance 5d4d97e1-5812-4884-a1ea-6eb0cee591c0 [ 900.561974] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355793, 'name': ReconfigVM_Task, 'duration_secs': 0.288848} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.562294] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 0c50937e-4646-47ec-ac0c-0281c9424d86/0c50937e-4646-47ec-ac0c-0281c9424d86.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 900.562923] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-44e36f18-c71a-4da9-be23-a8b10adeaaed {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.569438] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 900.569438] env[61970]: value = "task-1355795" [ 900.569438] env[61970]: _type = "Task" [ 900.569438] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.577577] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355795, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.591926] env[61970]: DEBUG oslo_vmware.api [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355794, 'name': ReconfigVM_Task, 'duration_secs': 0.105334} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.592497] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288764', 'volume_id': '4020284d-a273-459a-bb8c-84753ab98b1e', 'name': 'volume-4020284d-a273-459a-bb8c-84753ab98b1e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6838c6a4-8241-4b6f-ad58-3acfd7afdbf4', 'attached_at': '', 'detached_at': '', 'volume_id': '4020284d-a273-459a-bb8c-84753ab98b1e', 'serial': '4020284d-a273-459a-bb8c-84753ab98b1e'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 900.592497] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 900.593275] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-363c0dbe-8445-49a0-a8ac-b0ed59696b5f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.599768] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 900.600049] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-95047d85-311e-47ea-91bc-e3e84ec96683 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.686679] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 900.686934] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 900.687113] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 900.687296] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 900.687446] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 900.687594] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 900.687801] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 900.688123] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 900.688322] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 900.688494] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 900.688669] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 900.694048] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e40b4840-2185-461c-bd11-f776fc09fd4e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.714242] env[61970]: DEBUG oslo_vmware.api [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 900.714242] env[61970]: value = "task-1355797" [ 900.714242] env[61970]: _type = "Task" [ 900.714242] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.722994] env[61970]: DEBUG oslo_vmware.api [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355797, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.756447] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 900.756784] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 900.756986] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Deleting the datastore file [datastore1] 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.757300] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b16934c0-069e-4300-a842-e0819243f700 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.767504] env[61970]: DEBUG oslo_vmware.api [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Waiting for the task: (returnval){ [ 900.767504] env[61970]: value = "task-1355798" [ 900.767504] env[61970]: _type = "Task" [ 900.767504] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.780887] env[61970]: DEBUG oslo_vmware.api [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355798, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.958487] env[61970]: DEBUG oslo_concurrency.lockutils [None req-43749c25-99a2-4073-ae22-8e86ead4e847 tempest-ListServerFiltersTestJSON-1456607350 tempest-ListServerFiltersTestJSON-1456607350-project-member] Lock "5d4d97e1-5812-4884-a1ea-6eb0cee591c0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.256s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.002449] env[61970]: DEBUG nova.network.neutron [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Successfully updated port: b47e01b1-0321-4bbd-8ce3-d19568926d99 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 901.081849] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355795, 'name': Rename_Task, 'duration_secs': 0.138975} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.082369] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 901.082653] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a31ee2ef-3189-4202-bfc4-15a3799225de {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.090576] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 901.090576] env[61970]: value = "task-1355799" [ 901.090576] env[61970]: _type = "Task" [ 901.090576] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.099292] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355799, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.225882] env[61970]: DEBUG oslo_vmware.api [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355797, 'name': ReconfigVM_Task, 'duration_secs': 0.317267} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.226241] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Updating instance '75762ba3-f129-424e-a7cb-962785dfab2a' progress to 33 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 901.280942] env[61970]: DEBUG oslo_vmware.api [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Task: {'id': task-1355798, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186876} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.284186] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.284416] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 901.284619] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 901.284877] env[61970]: INFO nova.compute.manager [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Took 3.36 seconds to destroy the instance on the hypervisor. [ 901.285260] env[61970]: DEBUG oslo.service.loopingcall [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.286067] env[61970]: DEBUG nova.compute.manager [-] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 901.286172] env[61970]: DEBUG nova.network.neutron [-] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 901.510054] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.510654] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.517435] env[61970]: DEBUG nova.network.neutron [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 901.591305] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f44be5-d288-4ac1-8c16-5f1c87a8f0a1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.602957] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355799, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.605532] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1deedb9f-1eaf-4c22-8ff8-468f95523b0e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.636733] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-898d22dd-ad56-4059-b242-494d1e8a3816 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.644790] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d871377c-5dd8-4d23-ba2f-26384a8a4347 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.415398] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 902.415634] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 902.415783] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 902.415955] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 902.416104] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 902.416239] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 902.416441] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 902.416597] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 902.416752] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 902.416903] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 902.417073] env[61970]: DEBUG nova.virt.hardware [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 902.422173] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Reconfiguring VM instance instance-0000004b to detach disk 2000 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 902.425112] env[61970]: DEBUG nova.compute.manager [req-18142172-5d6d-4ba9-b14a-2e1b2fc5d9a0 req-ec93310d-ab7f-4dfa-863e-c433164f9b15 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Received event network-vif-plugged-b47e01b1-0321-4bbd-8ce3-d19568926d99 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.425509] env[61970]: DEBUG oslo_concurrency.lockutils [req-18142172-5d6d-4ba9-b14a-2e1b2fc5d9a0 req-ec93310d-ab7f-4dfa-863e-c433164f9b15 service nova] Acquiring lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.425710] env[61970]: DEBUG oslo_concurrency.lockutils [req-18142172-5d6d-4ba9-b14a-2e1b2fc5d9a0 req-ec93310d-ab7f-4dfa-863e-c433164f9b15 service nova] Lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.425871] env[61970]: DEBUG oslo_concurrency.lockutils [req-18142172-5d6d-4ba9-b14a-2e1b2fc5d9a0 req-ec93310d-ab7f-4dfa-863e-c433164f9b15 service nova] Lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.426039] env[61970]: DEBUG nova.compute.manager [req-18142172-5d6d-4ba9-b14a-2e1b2fc5d9a0 req-ec93310d-ab7f-4dfa-863e-c433164f9b15 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] No waiting events found dispatching network-vif-plugged-b47e01b1-0321-4bbd-8ce3-d19568926d99 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 902.426199] env[61970]: WARNING nova.compute.manager [req-18142172-5d6d-4ba9-b14a-2e1b2fc5d9a0 req-ec93310d-ab7f-4dfa-863e-c433164f9b15 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Received unexpected event network-vif-plugged-b47e01b1-0321-4bbd-8ce3-d19568926d99 for instance with vm_state active and task_state None. [ 902.426348] env[61970]: DEBUG nova.compute.manager [req-18142172-5d6d-4ba9-b14a-2e1b2fc5d9a0 req-ec93310d-ab7f-4dfa-863e-c433164f9b15 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Received event network-changed-b47e01b1-0321-4bbd-8ce3-d19568926d99 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.426512] env[61970]: DEBUG nova.compute.manager [req-18142172-5d6d-4ba9-b14a-2e1b2fc5d9a0 req-ec93310d-ab7f-4dfa-863e-c433164f9b15 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Refreshing instance network info cache due to event network-changed-b47e01b1-0321-4bbd-8ce3-d19568926d99. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 902.426678] env[61970]: DEBUG oslo_concurrency.lockutils [req-18142172-5d6d-4ba9-b14a-2e1b2fc5d9a0 req-ec93310d-ab7f-4dfa-863e-c433164f9b15 service nova] Acquiring lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.427200] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2068cef-8126-40fc-901e-77dfee6ca8e6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.453838] env[61970]: DEBUG nova.compute.provider_tree [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.459396] env[61970]: DEBUG oslo_vmware.api [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 902.459396] env[61970]: value = "task-1355800" [ 902.459396] env[61970]: _type = "Task" [ 902.459396] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.459543] env[61970]: DEBUG oslo_vmware.api [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355799, 'name': PowerOnVM_Task, 'duration_secs': 1.033036} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.460751] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 902.460751] env[61970]: INFO nova.compute.manager [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Took 9.56 seconds to spawn the instance on the hypervisor. [ 902.460751] env[61970]: DEBUG nova.compute.manager [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 902.464045] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-476953af-f174-40fe-aec0-9271eb181a66 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.477877] env[61970]: DEBUG oslo_vmware.api [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355800, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.519811] env[61970]: WARNING nova.network.neutron [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] 8f622fc5-b74f-4582-a6e0-44c90f495750 already exists in list: networks containing: ['8f622fc5-b74f-4582-a6e0-44c90f495750']. ignoring it [ 902.520759] env[61970]: WARNING nova.network.neutron [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] 8f622fc5-b74f-4582-a6e0-44c90f495750 already exists in list: networks containing: ['8f622fc5-b74f-4582-a6e0-44c90f495750']. ignoring it [ 902.962448] env[61970]: DEBUG nova.scheduler.client.report [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 902.983841] env[61970]: DEBUG oslo_vmware.api [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355800, 'name': ReconfigVM_Task, 'duration_secs': 0.167969} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.985839] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Reconfigured VM instance instance-0000004b to detach disk 2000 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 902.986613] env[61970]: INFO nova.compute.manager [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Took 20.65 seconds to build instance. [ 902.995106] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6956306c-0657-4a2c-8078-b532a033ddcb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.016523] env[61970]: DEBUG nova.network.neutron [-] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.026857] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 75762ba3-f129-424e-a7cb-962785dfab2a/75762ba3-f129-424e-a7cb-962785dfab2a.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 903.031170] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68860b4f-9c5e-4d65-91ce-729813546b4a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.059021] env[61970]: DEBUG oslo_vmware.api [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 903.059021] env[61970]: value = "task-1355801" [ 903.059021] env[61970]: _type = "Task" [ 903.059021] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.067412] env[61970]: DEBUG oslo_vmware.api [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355801, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.177821] env[61970]: DEBUG nova.network.neutron [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Updating instance_info_cache with network_info: [{"id": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "address": "fa:16:3e:14:8c:ec", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64a3e684-f5", "ovs_interfaceid": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9", "address": "fa:16:3e:45:a2:e3", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb33aa0c-3f", "ovs_interfaceid": "eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b47e01b1-0321-4bbd-8ce3-d19568926d99", "address": "fa:16:3e:32:95:cd", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb47e01b1-03", "ovs_interfaceid": "b47e01b1-0321-4bbd-8ce3-d19568926d99", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.470771] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.051s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.471474] env[61970]: DEBUG nova.compute.manager [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 903.474592] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.120s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.474878] env[61970]: DEBUG nova.objects.instance [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Lazy-loading 'resources' on Instance uuid b385407b-1bdd-4c53-907c-cb4c8ce16cc7 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.500871] env[61970]: DEBUG oslo_concurrency.lockutils [None req-45d87880-3517-49de-8157-422e9e04c7da tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "0c50937e-4646-47ec-ac0c-0281c9424d86" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.175s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.528384] env[61970]: INFO nova.compute.manager [-] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Took 2.24 seconds to deallocate network for instance. [ 903.567872] env[61970]: DEBUG oslo_vmware.api [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355801, 'name': ReconfigVM_Task, 'duration_secs': 0.400705} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.568435] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 75762ba3-f129-424e-a7cb-962785dfab2a/75762ba3-f129-424e-a7cb-962785dfab2a.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 903.568738] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Updating instance '75762ba3-f129-424e-a7cb-962785dfab2a' progress to 50 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 903.681208] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.681904] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.682074] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.682357] env[61970]: DEBUG oslo_concurrency.lockutils [req-18142172-5d6d-4ba9-b14a-2e1b2fc5d9a0 req-ec93310d-ab7f-4dfa-863e-c433164f9b15 service nova] Acquired lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.682537] env[61970]: DEBUG nova.network.neutron [req-18142172-5d6d-4ba9-b14a-2e1b2fc5d9a0 req-ec93310d-ab7f-4dfa-863e-c433164f9b15 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Refreshing network info cache for port b47e01b1-0321-4bbd-8ce3-d19568926d99 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 903.684742] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448c80e3-8219-47df-b997-16c448c96bcf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.703165] env[61970]: DEBUG nova.virt.hardware [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 903.703442] env[61970]: DEBUG nova.virt.hardware [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 903.703600] env[61970]: DEBUG nova.virt.hardware [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 903.703785] env[61970]: DEBUG nova.virt.hardware [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 903.703926] env[61970]: DEBUG nova.virt.hardware [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 903.704087] env[61970]: DEBUG nova.virt.hardware [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 903.704298] env[61970]: DEBUG nova.virt.hardware [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 903.704984] env[61970]: DEBUG nova.virt.hardware [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 903.704984] env[61970]: DEBUG nova.virt.hardware [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 903.704984] env[61970]: DEBUG nova.virt.hardware [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 903.705204] env[61970]: DEBUG nova.virt.hardware [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 903.711387] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Reconfiguring VM to attach interface {{(pid=61970) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 903.712307] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b78ad28-675b-44bc-90af-0a5219ce7243 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.731391] env[61970]: DEBUG oslo_vmware.api [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 903.731391] env[61970]: value = "task-1355802" [ 903.731391] env[61970]: _type = "Task" [ 903.731391] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.740368] env[61970]: DEBUG oslo_vmware.api [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355802, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.978583] env[61970]: DEBUG nova.compute.utils [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 903.980600] env[61970]: DEBUG nova.compute.manager [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 903.980600] env[61970]: DEBUG nova.network.neutron [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 904.017089] env[61970]: DEBUG oslo_concurrency.lockutils [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "0c50937e-4646-47ec-ac0c-0281c9424d86" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.017391] env[61970]: DEBUG oslo_concurrency.lockutils [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "0c50937e-4646-47ec-ac0c-0281c9424d86" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.017686] env[61970]: INFO nova.compute.manager [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Shelving [ 904.023938] env[61970]: DEBUG nova.compute.manager [req-fdba64ef-3e4b-4c64-82ab-00d708f4f4af req-b398d872-63e3-4fc2-827e-0ab532028123 service nova] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Received event network-vif-deleted-1dfb626c-a7e0-455e-bffd-3eabada94ae1 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.075390] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84f71537-7c34-44f4-9465-612fd85311aa {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.079953] env[61970]: DEBUG nova.policy [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf68cfc72ee04882ad02b827751520bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1d4724cbb0841859369af01f2c03db1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 904.108236] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b6be3a-5fd3-48be-8bad-e8fe0ade0ff9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.112209] env[61970]: INFO nova.compute.manager [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Took 0.58 seconds to detach 1 volumes for instance. [ 904.115951] env[61970]: DEBUG nova.compute.manager [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Deleting volume: 4020284d-a273-459a-bb8c-84753ab98b1e {{(pid=61970) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 904.131413] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Updating instance '75762ba3-f129-424e-a7cb-962785dfab2a' progress to 67 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 904.211455] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a0101bd-bf8a-4b21-a41a-2d840a52b3e2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.225893] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f98dd0c5-5bb4-4212-92eb-d0750cf39c93 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.244185] env[61970]: DEBUG oslo_vmware.api [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355802, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.273661] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3611c256-16a0-473c-bb91-1ebaaedbefa3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.283353] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c803290-0186-4e63-87c0-f66cb97871e1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.296149] env[61970]: DEBUG nova.compute.provider_tree [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.486904] env[61970]: DEBUG nova.compute.manager [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 904.529966] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 904.530271] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e4d2a61f-36d8-46de-82d7-b774f6e56e7d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.541425] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 904.541425] env[61970]: value = "task-1355804" [ 904.541425] env[61970]: _type = "Task" [ 904.541425] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.550919] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355804, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.576425] env[61970]: DEBUG nova.network.neutron [req-18142172-5d6d-4ba9-b14a-2e1b2fc5d9a0 req-ec93310d-ab7f-4dfa-863e-c433164f9b15 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Updated VIF entry in instance network info cache for port b47e01b1-0321-4bbd-8ce3-d19568926d99. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 904.577126] env[61970]: DEBUG nova.network.neutron [req-18142172-5d6d-4ba9-b14a-2e1b2fc5d9a0 req-ec93310d-ab7f-4dfa-863e-c433164f9b15 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Updating instance_info_cache with network_info: [{"id": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "address": "fa:16:3e:14:8c:ec", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64a3e684-f5", "ovs_interfaceid": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9", "address": "fa:16:3e:45:a2:e3", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb33aa0c-3f", "ovs_interfaceid": "eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b47e01b1-0321-4bbd-8ce3-d19568926d99", "address": "fa:16:3e:32:95:cd", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb47e01b1-03", "ovs_interfaceid": "b47e01b1-0321-4bbd-8ce3-d19568926d99", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.586702] env[61970]: DEBUG nova.network.neutron [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Successfully created port: 5c475b00-f4ed-4e09-916c-6f3ca7845784 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 904.675499] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.713937] env[61970]: DEBUG nova.network.neutron [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Port 32711b6a-a9cc-472d-a634-e3b60711bffe binding to destination host cpu-1 is already ACTIVE {{(pid=61970) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 904.757245] env[61970]: DEBUG oslo_vmware.api [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355802, 'name': ReconfigVM_Task, 'duration_secs': 0.986826} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.757865] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.758148] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Reconfigured VM to attach interface {{(pid=61970) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 904.801346] env[61970]: DEBUG nova.scheduler.client.report [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.054178] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355804, 'name': PowerOffVM_Task, 'duration_secs': 0.25552} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.054815] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 905.055815] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa3ba20e-14e6-4f34-9b7f-ec0bce95d38f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.081032] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58af9ff8-94b2-49ff-bba6-251db1eb27c1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.083907] env[61970]: DEBUG oslo_concurrency.lockutils [req-18142172-5d6d-4ba9-b14a-2e1b2fc5d9a0 req-ec93310d-ab7f-4dfa-863e-c433164f9b15 service nova] Releasing lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.263957] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8e9f5e6d-5c0a-4fdc-a93e-128c15decfa8 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "interface-b1cc1cc2-15d6-459d-9529-e592ddb225ac-b47e01b1-0321-4bbd-8ce3-d19568926d99" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.570s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.306897] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.832s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.309467] env[61970]: DEBUG oslo_concurrency.lockutils [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.935s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.309714] env[61970]: DEBUG nova.objects.instance [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lazy-loading 'resources' on Instance uuid 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 905.333045] env[61970]: INFO nova.scheduler.client.report [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Deleted allocations for instance b385407b-1bdd-4c53-907c-cb4c8ce16cc7 [ 905.501726] env[61970]: DEBUG nova.compute.manager [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 905.530969] env[61970]: DEBUG nova.virt.hardware [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 905.531344] env[61970]: DEBUG nova.virt.hardware [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 905.531504] env[61970]: DEBUG nova.virt.hardware [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 905.531686] env[61970]: DEBUG nova.virt.hardware [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 905.531828] env[61970]: DEBUG nova.virt.hardware [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 905.531973] env[61970]: DEBUG nova.virt.hardware [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 905.532198] env[61970]: DEBUG nova.virt.hardware [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 905.532355] env[61970]: DEBUG nova.virt.hardware [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 905.532528] env[61970]: DEBUG nova.virt.hardware [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 905.532692] env[61970]: DEBUG nova.virt.hardware [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 905.532863] env[61970]: DEBUG nova.virt.hardware [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 905.533783] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4329a55-d414-4d55-bc37-7addd7cee64e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.542735] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84562378-b8b5-4f25-842b-b8f888b50044 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.594228] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Creating Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 905.594564] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1a986afc-a251-491f-9c44-766c9a1ad6bd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.603703] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 905.603703] env[61970]: value = "task-1355805" [ 905.603703] env[61970]: _type = "Task" [ 905.603703] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.612699] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355805, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.749079] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "75762ba3-f129-424e-a7cb-962785dfab2a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.749323] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "75762ba3-f129-424e-a7cb-962785dfab2a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.749507] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "75762ba3-f129-424e-a7cb-962785dfab2a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.844791] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7a4acf10-14b3-4b67-b1bb-d39e6b9465ab tempest-ServersTestManualDisk-412229977 tempest-ServersTestManualDisk-412229977-project-member] Lock "b385407b-1bdd-4c53-907c-cb4c8ce16cc7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.890s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.960144] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "ebb92a5b-f635-4504-a57f-395de514015a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.960342] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "ebb92a5b-f635-4504-a57f-395de514015a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.982818] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91572763-9a53-4e21-8398-63b6c84b9115 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.992812] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c34efd3-107d-493d-85e9-53d191c0877a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.030793] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75b00311-4f36-4dba-a7b7-c25030832f3c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.039879] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a456d61d-6e8d-4b1c-92aa-c04d5d4dbd90 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.055323] env[61970]: DEBUG nova.compute.provider_tree [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.117322] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355805, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.256855] env[61970]: DEBUG nova.compute.manager [req-85e31945-811e-4f09-b2d6-502c5d2366ee req-548312bb-ac3f-4373-a645-d219d1448516 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Received event network-vif-plugged-5c475b00-f4ed-4e09-916c-6f3ca7845784 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.256964] env[61970]: DEBUG oslo_concurrency.lockutils [req-85e31945-811e-4f09-b2d6-502c5d2366ee req-548312bb-ac3f-4373-a645-d219d1448516 service nova] Acquiring lock "7f59000a-94d9-45b6-aa7a-300d95793615-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.258749] env[61970]: DEBUG oslo_concurrency.lockutils [req-85e31945-811e-4f09-b2d6-502c5d2366ee req-548312bb-ac3f-4373-a645-d219d1448516 service nova] Lock "7f59000a-94d9-45b6-aa7a-300d95793615-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.258749] env[61970]: DEBUG oslo_concurrency.lockutils [req-85e31945-811e-4f09-b2d6-502c5d2366ee req-548312bb-ac3f-4373-a645-d219d1448516 service nova] Lock "7f59000a-94d9-45b6-aa7a-300d95793615-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.258749] env[61970]: DEBUG nova.compute.manager [req-85e31945-811e-4f09-b2d6-502c5d2366ee req-548312bb-ac3f-4373-a645-d219d1448516 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] No waiting events found dispatching network-vif-plugged-5c475b00-f4ed-4e09-916c-6f3ca7845784 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 906.258749] env[61970]: WARNING nova.compute.manager [req-85e31945-811e-4f09-b2d6-502c5d2366ee req-548312bb-ac3f-4373-a645-d219d1448516 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Received unexpected event network-vif-plugged-5c475b00-f4ed-4e09-916c-6f3ca7845784 for instance with vm_state building and task_state spawning. [ 906.349689] env[61970]: DEBUG nova.network.neutron [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Successfully updated port: 5c475b00-f4ed-4e09-916c-6f3ca7845784 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 906.467629] env[61970]: DEBUG nova.compute.manager [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 906.560793] env[61970]: DEBUG nova.scheduler.client.report [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 906.617021] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355805, 'name': CreateSnapshot_Task, 'duration_secs': 0.595557} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.617291] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Created Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 906.618082] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-269980a6-2407-4f5f-8d81-5914d519873c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.805570] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "refresh_cache-75762ba3-f129-424e-a7cb-962785dfab2a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.805728] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "refresh_cache-75762ba3-f129-424e-a7cb-962785dfab2a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.805900] env[61970]: DEBUG nova.network.neutron [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 906.851048] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.851279] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquired lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.851472] env[61970]: DEBUG nova.network.neutron [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 906.991650] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.066120] env[61970]: DEBUG oslo_concurrency.lockutils [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.757s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.068456] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.396s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.069168] env[61970]: DEBUG nova.objects.instance [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Lazy-loading 'resources' on Instance uuid 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.094945] env[61970]: INFO nova.scheduler.client.report [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleted allocations for instance 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20 [ 907.141982] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Creating linked-clone VM from snapshot {{(pid=61970) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 907.144032] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-36c64b28-503e-4ef3-956c-6c2e7c84df2a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.153584] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "interface-b1cc1cc2-15d6-459d-9529-e592ddb225ac-eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.153841] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "interface-b1cc1cc2-15d6-459d-9529-e592ddb225ac-eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.155157] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 907.155157] env[61970]: value = "task-1355806" [ 907.155157] env[61970]: _type = "Task" [ 907.155157] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.165781] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355806, 'name': CloneVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.412008] env[61970]: DEBUG nova.network.neutron [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 907.603756] env[61970]: DEBUG oslo_concurrency.lockutils [None req-29f920cb-e345-4351-b21b-dce16b68588d tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.524s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.662995] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.663210] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.664272] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28948858-2f7e-4164-b8af-45515b081476 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.673312] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355806, 'name': CloneVM_Task} progress is 94%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.690578] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35b07505-add0-4fdd-823b-31f6d0396dfc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.722801] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Reconfiguring VM to detach interface {{(pid=61970) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 907.725801] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b693d1e-ac7a-4b35-a45b-8e92e2b31359 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.758805] env[61970]: DEBUG oslo_vmware.api [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 907.758805] env[61970]: value = "task-1355807" [ 907.758805] env[61970]: _type = "Task" [ 907.758805] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.774461] env[61970]: DEBUG oslo_vmware.api [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355807, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.798580] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db000565-041d-4ce9-a873-c050b2687bd0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.808032] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19e916c1-4aef-4ef2-b1f0-32112acb575c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.841508] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c5d71fe-390f-40e5-ac25-4912b3c4c390 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.850174] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef624d5-1e48-4dee-9cca-a658f3eecfb9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.864630] env[61970]: DEBUG nova.compute.provider_tree [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.894635] env[61970]: DEBUG nova.network.neutron [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Updating instance_info_cache with network_info: [{"id": "5c475b00-f4ed-4e09-916c-6f3ca7845784", "address": "fa:16:3e:27:f3:40", "network": {"id": "70081995-3d6c-466e-b811-1c8e532e552f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-641102750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1d4724cbb0841859369af01f2c03db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aaf1b231-3660-4453-b4f3-44d825b9a5dd", "external-id": "nsx-vlan-transportzone-6", "segmentation_id": 6, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c475b00-f4", "ovs_interfaceid": "5c475b00-f4ed-4e09-916c-6f3ca7845784", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.898896] env[61970]: DEBUG nova.network.neutron [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Updating instance_info_cache with network_info: [{"id": "32711b6a-a9cc-472d-a634-e3b60711bffe", "address": "fa:16:3e:68:3d:bc", "network": {"id": "cba185ff-bfa3-4547-b7b0-638c3e79a24c", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-864852491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5c64cb9232fb413cbd7627dcf077e9ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32711b6a-a9", "ovs_interfaceid": "32711b6a-a9cc-472d-a634-e3b60711bffe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.167692] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355806, 'name': CloneVM_Task} progress is 95%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.269887] env[61970]: DEBUG oslo_vmware.api [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355807, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.282171] env[61970]: DEBUG nova.compute.manager [req-d4a78712-3769-4e51-bf02-75d6d92dbef4 req-b19236bd-c1d2-4af9-ac8d-24a0d36c7154 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Received event network-changed-5c475b00-f4ed-4e09-916c-6f3ca7845784 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.282320] env[61970]: DEBUG nova.compute.manager [req-d4a78712-3769-4e51-bf02-75d6d92dbef4 req-b19236bd-c1d2-4af9-ac8d-24a0d36c7154 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Refreshing instance network info cache due to event network-changed-5c475b00-f4ed-4e09-916c-6f3ca7845784. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 908.282545] env[61970]: DEBUG oslo_concurrency.lockutils [req-d4a78712-3769-4e51-bf02-75d6d92dbef4 req-b19236bd-c1d2-4af9-ac8d-24a0d36c7154 service nova] Acquiring lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.368894] env[61970]: DEBUG nova.scheduler.client.report [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 908.391174] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "3b346c68-53a9-4ceb-83a3-9e4bce9610b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.391412] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "3b346c68-53a9-4ceb-83a3-9e4bce9610b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.397432] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Releasing lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.397724] env[61970]: DEBUG nova.compute.manager [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Instance network_info: |[{"id": "5c475b00-f4ed-4e09-916c-6f3ca7845784", "address": "fa:16:3e:27:f3:40", "network": {"id": "70081995-3d6c-466e-b811-1c8e532e552f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-641102750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1d4724cbb0841859369af01f2c03db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aaf1b231-3660-4453-b4f3-44d825b9a5dd", "external-id": "nsx-vlan-transportzone-6", "segmentation_id": 6, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c475b00-f4", "ovs_interfaceid": "5c475b00-f4ed-4e09-916c-6f3ca7845784", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 908.398232] env[61970]: DEBUG oslo_concurrency.lockutils [req-d4a78712-3769-4e51-bf02-75d6d92dbef4 req-b19236bd-c1d2-4af9-ac8d-24a0d36c7154 service nova] Acquired lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.398418] env[61970]: DEBUG nova.network.neutron [req-d4a78712-3769-4e51-bf02-75d6d92dbef4 req-b19236bd-c1d2-4af9-ac8d-24a0d36c7154 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Refreshing network info cache for port 5c475b00-f4ed-4e09-916c-6f3ca7845784 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 908.400300] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:f3:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aaf1b231-3660-4453-b4f3-44d825b9a5dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5c475b00-f4ed-4e09-916c-6f3ca7845784', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 908.407712] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Creating folder: Project (e1d4724cbb0841859369af01f2c03db1). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 908.408996] env[61970]: DEBUG oslo_concurrency.lockutils [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "refresh_cache-75762ba3-f129-424e-a7cb-962785dfab2a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.412549] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f98128ea-712e-4717-a3e6-7d29acef9e57 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.426625] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Created folder: Project (e1d4724cbb0841859369af01f2c03db1) in parent group-v288740. [ 908.426904] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Creating folder: Instances. Parent ref: group-v288823. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 908.427165] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8585a91b-ad7e-4153-a75d-53e973e93941 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.438206] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Created folder: Instances in parent group-v288823. [ 908.438206] env[61970]: DEBUG oslo.service.loopingcall [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 908.438366] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 908.438523] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e1c8dc6-7d20-4a56-a4a4-11c1ceddf080 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.459771] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 908.459771] env[61970]: value = "task-1355810" [ 908.459771] env[61970]: _type = "Task" [ 908.459771] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.470156] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355810, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.668274] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355806, 'name': CloneVM_Task, 'duration_secs': 1.178353} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.668486] env[61970]: INFO nova.virt.vmwareapi.vmops [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Created linked-clone VM from snapshot [ 908.669264] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9997871-5978-40da-89c8-803e8780ba66 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.678748] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Uploading image c0a6b383-5802-40b2-b2d3-afa9fec0129b {{(pid=61970) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 908.707997] env[61970]: DEBUG oslo_vmware.rw_handles [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 908.707997] env[61970]: value = "vm-288822" [ 908.707997] env[61970]: _type = "VirtualMachine" [ 908.707997] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 908.708376] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-190dd453-b133-4231-90fd-eaf8e9e17c39 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.716645] env[61970]: DEBUG oslo_vmware.rw_handles [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lease: (returnval){ [ 908.716645] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52720eb3-40d0-801d-39b8-e442fd916e34" [ 908.716645] env[61970]: _type = "HttpNfcLease" [ 908.716645] env[61970]: } obtained for exporting VM: (result){ [ 908.716645] env[61970]: value = "vm-288822" [ 908.716645] env[61970]: _type = "VirtualMachine" [ 908.716645] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 908.717250] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the lease: (returnval){ [ 908.717250] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52720eb3-40d0-801d-39b8-e442fd916e34" [ 908.717250] env[61970]: _type = "HttpNfcLease" [ 908.717250] env[61970]: } to be ready. {{(pid=61970) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 908.724329] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 908.724329] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52720eb3-40d0-801d-39b8-e442fd916e34" [ 908.724329] env[61970]: _type = "HttpNfcLease" [ 908.724329] env[61970]: } is initializing. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 908.770157] env[61970]: DEBUG oslo_vmware.api [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355807, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.873606] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.805s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.875859] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.884s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.877470] env[61970]: INFO nova.compute.claims [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 908.894222] env[61970]: DEBUG nova.compute.manager [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 908.897513] env[61970]: INFO nova.scheduler.client.report [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Deleted allocations for instance 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4 [ 908.930884] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfde5840-68be-4653-810c-b35ea9f606c0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.955910] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1de7b026-8393-4c5f-8569-47fd9beb3f46 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.967408] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Updating instance '75762ba3-f129-424e-a7cb-962785dfab2a' progress to 83 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 908.976968] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355810, 'name': CreateVM_Task, 'duration_secs': 0.365469} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.976968] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 908.977626] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.977803] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.978151] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 908.978408] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82d44412-e1de-4bb3-b409-9519b56b2d7b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.983643] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 908.983643] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523762ea-e8e3-cdc8-903d-8da1569b94a8" [ 908.983643] env[61970]: _type = "Task" [ 908.983643] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.992324] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523762ea-e8e3-cdc8-903d-8da1569b94a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.186186] env[61970]: DEBUG nova.network.neutron [req-d4a78712-3769-4e51-bf02-75d6d92dbef4 req-b19236bd-c1d2-4af9-ac8d-24a0d36c7154 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Updated VIF entry in instance network info cache for port 5c475b00-f4ed-4e09-916c-6f3ca7845784. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 909.186595] env[61970]: DEBUG nova.network.neutron [req-d4a78712-3769-4e51-bf02-75d6d92dbef4 req-b19236bd-c1d2-4af9-ac8d-24a0d36c7154 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Updating instance_info_cache with network_info: [{"id": "5c475b00-f4ed-4e09-916c-6f3ca7845784", "address": "fa:16:3e:27:f3:40", "network": {"id": "70081995-3d6c-466e-b811-1c8e532e552f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-641102750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1d4724cbb0841859369af01f2c03db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aaf1b231-3660-4453-b4f3-44d825b9a5dd", "external-id": "nsx-vlan-transportzone-6", "segmentation_id": 6, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c475b00-f4", "ovs_interfaceid": "5c475b00-f4ed-4e09-916c-6f3ca7845784", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.227032] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 909.227032] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52720eb3-40d0-801d-39b8-e442fd916e34" [ 909.227032] env[61970]: _type = "HttpNfcLease" [ 909.227032] env[61970]: } is ready. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 909.227385] env[61970]: DEBUG oslo_vmware.rw_handles [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 909.227385] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52720eb3-40d0-801d-39b8-e442fd916e34" [ 909.227385] env[61970]: _type = "HttpNfcLease" [ 909.227385] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 909.228166] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ba8f798-f7cb-4b7a-a86e-83dc98c71e69 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.238555] env[61970]: DEBUG oslo_vmware.rw_handles [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52301958-2576-ce1d-b115-9973e2019ca1/disk-0.vmdk from lease info. {{(pid=61970) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 909.238778] env[61970]: DEBUG oslo_vmware.rw_handles [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52301958-2576-ce1d-b115-9973e2019ca1/disk-0.vmdk for reading. {{(pid=61970) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 909.309105] env[61970]: DEBUG oslo_vmware.api [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355807, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.331632] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-cbb94f04-f073-4c99-8edd-1d5ee10de558 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.405837] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f3e463b0-8204-4949-93b0-a2353024bba6 tempest-ServersTestBootFromVolume-687806 tempest-ServersTestBootFromVolume-687806-project-member] Lock "6838c6a4-8241-4b6f-ad58-3acfd7afdbf4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.486s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.423769] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.474332] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 909.474644] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1681d38b-d728-4efe-b9eb-05e6a657c252 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.483854] env[61970]: DEBUG oslo_vmware.api [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 909.483854] env[61970]: value = "task-1355812" [ 909.483854] env[61970]: _type = "Task" [ 909.483854] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.498759] env[61970]: DEBUG oslo_vmware.api [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355812, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.504391] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523762ea-e8e3-cdc8-903d-8da1569b94a8, 'name': SearchDatastore_Task, 'duration_secs': 0.021688} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.505707] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.505707] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 909.505707] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.505707] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.505947] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 909.506384] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-02878709-02fd-4f3e-9c20-0eea8347e5e4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.524972] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 909.525179] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 909.525954] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c966f9ee-f006-4b05-8184-baaaa1f48b47 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.532799] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 909.532799] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52800d94-a995-b7e2-9424-80ee4e20391d" [ 909.532799] env[61970]: _type = "Task" [ 909.532799] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.541546] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52800d94-a995-b7e2-9424-80ee4e20391d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.689479] env[61970]: DEBUG oslo_concurrency.lockutils [req-d4a78712-3769-4e51-bf02-75d6d92dbef4 req-b19236bd-c1d2-4af9-ac8d-24a0d36c7154 service nova] Releasing lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.809367] env[61970]: DEBUG oslo_vmware.api [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355807, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.995817] env[61970]: DEBUG oslo_vmware.api [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355812, 'name': PowerOnVM_Task, 'duration_secs': 0.436359} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.996411] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 909.996411] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-85f23552-ac1a-4c56-9dd8-1e33c7ee67be tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Updating instance '75762ba3-f129-424e-a7cb-962785dfab2a' progress to 100 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 910.049440] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52800d94-a995-b7e2-9424-80ee4e20391d, 'name': SearchDatastore_Task, 'duration_secs': 0.009802} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.052899] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-340bb40b-0635-4063-b843-1ea897ac2cc5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.064752] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 910.064752] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f88500-f835-bb52-5c13-edbbbf5128bc" [ 910.064752] env[61970]: _type = "Task" [ 910.064752] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.077262] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f88500-f835-bb52-5c13-edbbbf5128bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.092892] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd2182ee-fdbe-4b89-a456-7d3983fe7b61 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.104511] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2afb4e50-03c1-4a70-a10b-48a5b18f71de {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.143532] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-134dee35-2554-4e6d-8f8e-265b36bb9884 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.153109] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3648ff9-bb30-4c64-9a0f-9f21b28db3b4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.170022] env[61970]: DEBUG nova.compute.provider_tree [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.311481] env[61970]: DEBUG oslo_vmware.api [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355807, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.576656] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f88500-f835-bb52-5c13-edbbbf5128bc, 'name': SearchDatastore_Task, 'duration_secs': 0.018595} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.576975] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.577262] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 7f59000a-94d9-45b6-aa7a-300d95793615/7f59000a-94d9-45b6-aa7a-300d95793615.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 910.577535] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-57019604-b007-4eac-ac29-cf24ee8114ed {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.585834] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 910.585834] env[61970]: value = "task-1355813" [ 910.585834] env[61970]: _type = "Task" [ 910.585834] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.595163] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355813, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.674407] env[61970]: DEBUG nova.scheduler.client.report [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 910.812412] env[61970]: DEBUG oslo_vmware.api [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355807, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.097967] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355813, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.185055] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.308s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.185055] env[61970]: DEBUG nova.compute.manager [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 911.187846] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.764s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.189537] env[61970]: INFO nova.compute.claims [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 911.310178] env[61970]: DEBUG oslo_vmware.api [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355807, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.557532] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.557978] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.598091] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355813, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.64013} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.598447] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 7f59000a-94d9-45b6-aa7a-300d95793615/7f59000a-94d9-45b6-aa7a-300d95793615.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 911.598699] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 911.599017] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-184849c6-eaad-427f-91c1-f3288ef92fbf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.608315] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 911.608315] env[61970]: value = "task-1355814" [ 911.608315] env[61970]: _type = "Task" [ 911.608315] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.618119] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355814, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.694474] env[61970]: DEBUG nova.compute.utils [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 911.699191] env[61970]: DEBUG nova.compute.manager [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 911.699428] env[61970]: DEBUG nova.network.neutron [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 911.742881] env[61970]: DEBUG nova.policy [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2b8322c5de6483aaf032781d28e950d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09033ee688384ca287b4fdad6e67cb1f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 911.810754] env[61970]: DEBUG oslo_vmware.api [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355807, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.035885] env[61970]: DEBUG nova.network.neutron [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Successfully created port: 9a1c8b46-0249-4e0b-94de-b0d0062ca69c {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 912.061873] env[61970]: DEBUG nova.compute.utils [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 912.119903] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355814, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092531} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.120200] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 912.120998] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21269b08-0696-40ac-9910-3310d5597bfc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.144883] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 7f59000a-94d9-45b6-aa7a-300d95793615/7f59000a-94d9-45b6-aa7a-300d95793615.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 912.147349] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-15db4750-8e55-46f1-9ef0-4cc4db0a8472 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.171684] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 912.171684] env[61970]: value = "task-1355815" [ 912.171684] env[61970]: _type = "Task" [ 912.171684] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.182398] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355815, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.200300] env[61970]: DEBUG nova.compute.manager [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 912.314568] env[61970]: DEBUG oslo_vmware.api [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355807, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.369018] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577a935a-7ef9-4dd7-b37f-1f51e8250fe3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.371066] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "75762ba3-f129-424e-a7cb-962785dfab2a" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.371475] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "75762ba3-f129-424e-a7cb-962785dfab2a" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.371811] env[61970]: DEBUG nova.compute.manager [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Going to confirm migration 1 {{(pid=61970) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 912.378841] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b372cb89-dc16-4c12-8be7-34227bd86962 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.411923] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75399804-bc2c-4208-8e97-02a7f3e81b7d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.420770] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2667c4b-cb69-4d9e-95c6-c35db4b05e31 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.436299] env[61970]: DEBUG nova.compute.provider_tree [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 912.566818] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.683221] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355815, 'name': ReconfigVM_Task, 'duration_secs': 0.502349} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.683747] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 7f59000a-94d9-45b6-aa7a-300d95793615/7f59000a-94d9-45b6-aa7a-300d95793615.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.684547] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e1cd08b1-fa6e-44ae-bc38-e4b38808ba06 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.694378] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 912.694378] env[61970]: value = "task-1355816" [ 912.694378] env[61970]: _type = "Task" [ 912.694378] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.705413] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355816, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.814632] env[61970]: DEBUG oslo_vmware.api [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355807, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.957702] env[61970]: ERROR nova.scheduler.client.report [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [req-0e2fcbf4-0343-494d-bca9-ea2b92ab199b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fc0f3e4e-5ef5-45e9-8335-cc68743405aa. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-0e2fcbf4-0343-494d-bca9-ea2b92ab199b"}]} [ 912.977764] env[61970]: DEBUG nova.scheduler.client.report [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Refreshing inventories for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 912.990177] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "refresh_cache-75762ba3-f129-424e-a7cb-962785dfab2a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.990551] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "refresh_cache-75762ba3-f129-424e-a7cb-962785dfab2a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.990769] env[61970]: DEBUG nova.network.neutron [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 912.990989] env[61970]: DEBUG nova.objects.instance [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lazy-loading 'info_cache' on Instance uuid 75762ba3-f129-424e-a7cb-962785dfab2a {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.997616] env[61970]: DEBUG nova.scheduler.client.report [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Updating ProviderTree inventory for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 912.997955] env[61970]: DEBUG nova.compute.provider_tree [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 913.012858] env[61970]: DEBUG nova.scheduler.client.report [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Refreshing aggregate associations for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa, aggregates: None {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 913.033747] env[61970]: DEBUG nova.scheduler.client.report [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Refreshing trait associations for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 913.204627] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355816, 'name': Rename_Task, 'duration_secs': 0.284098} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.207543] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 913.208203] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f71eb205-3ad1-4f53-8c2e-565658448f94 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.214416] env[61970]: DEBUG nova.compute.manager [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 913.218294] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 913.218294] env[61970]: value = "task-1355817" [ 913.218294] env[61970]: _type = "Task" [ 913.218294] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.225195] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ea93bc-6423-49b4-812e-9cb27eca5a12 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.234975] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355817, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.238745] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c1602e-ad8f-4a0a-be64-a3dd4f11b194 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.286495] env[61970]: DEBUG nova.virt.hardware [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 913.287054] env[61970]: DEBUG nova.virt.hardware [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 913.287054] env[61970]: DEBUG nova.virt.hardware [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 913.287293] env[61970]: DEBUG nova.virt.hardware [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 913.287357] env[61970]: DEBUG nova.virt.hardware [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 913.287521] env[61970]: DEBUG nova.virt.hardware [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 913.287739] env[61970]: DEBUG nova.virt.hardware [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 913.287921] env[61970]: DEBUG nova.virt.hardware [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 913.288269] env[61970]: DEBUG nova.virt.hardware [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 913.288339] env[61970]: DEBUG nova.virt.hardware [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 913.288496] env[61970]: DEBUG nova.virt.hardware [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 913.289426] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ce6ec5-e80e-4a8f-95f4-a22b33593477 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.292913] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf56c30-c839-40ee-8f2d-6992400efc96 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.305370] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae4abdd-9f55-46e9-931e-faa718d08039 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.311042] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb76118-6119-4569-9bd4-ac5db3544dfd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.337978] env[61970]: DEBUG nova.compute.provider_tree [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 913.342675] env[61970]: DEBUG oslo_vmware.api [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355807, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.646717] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.647062] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.648373] env[61970]: INFO nova.compute.manager [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Attaching volume 148b4432-023f-47fd-b4c0-7e51f31edf89 to /dev/sdb [ 913.666279] env[61970]: DEBUG nova.compute.manager [req-e0747183-ca01-4aad-a678-603793eb5967 req-3b2d1595-e729-43c2-9201-d84b6a31b4d0 service nova] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Received event network-vif-plugged-9a1c8b46-0249-4e0b-94de-b0d0062ca69c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.667292] env[61970]: DEBUG oslo_concurrency.lockutils [req-e0747183-ca01-4aad-a678-603793eb5967 req-3b2d1595-e729-43c2-9201-d84b6a31b4d0 service nova] Acquiring lock "ebb92a5b-f635-4504-a57f-395de514015a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.667292] env[61970]: DEBUG oslo_concurrency.lockutils [req-e0747183-ca01-4aad-a678-603793eb5967 req-3b2d1595-e729-43c2-9201-d84b6a31b4d0 service nova] Lock "ebb92a5b-f635-4504-a57f-395de514015a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.667292] env[61970]: DEBUG oslo_concurrency.lockutils [req-e0747183-ca01-4aad-a678-603793eb5967 req-3b2d1595-e729-43c2-9201-d84b6a31b4d0 service nova] Lock "ebb92a5b-f635-4504-a57f-395de514015a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.667292] env[61970]: DEBUG nova.compute.manager [req-e0747183-ca01-4aad-a678-603793eb5967 req-3b2d1595-e729-43c2-9201-d84b6a31b4d0 service nova] [instance: ebb92a5b-f635-4504-a57f-395de514015a] No waiting events found dispatching network-vif-plugged-9a1c8b46-0249-4e0b-94de-b0d0062ca69c {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 913.667500] env[61970]: WARNING nova.compute.manager [req-e0747183-ca01-4aad-a678-603793eb5967 req-3b2d1595-e729-43c2-9201-d84b6a31b4d0 service nova] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Received unexpected event network-vif-plugged-9a1c8b46-0249-4e0b-94de-b0d0062ca69c for instance with vm_state building and task_state spawning. [ 913.694095] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7ca24c-bb93-42d5-a70c-54d48bcbd41b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.702722] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-379e53a5-9053-4ec6-b1a5-932f122aea53 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.719396] env[61970]: DEBUG nova.virt.block_device [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Updating existing volume attachment record: be240b2e-7ee6-4768-9a95-9094355ab8e3 {{(pid=61970) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 913.734549] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355817, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.821288] env[61970]: DEBUG oslo_vmware.api [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355807, 'name': ReconfigVM_Task, 'duration_secs': 5.795241} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.821612] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.821816] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Reconfigured VM to detach interface {{(pid=61970) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 913.878309] env[61970]: DEBUG nova.scheduler.client.report [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Updated inventory for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with generation 97 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 913.878676] env[61970]: DEBUG nova.compute.provider_tree [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Updating resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa generation from 97 to 98 during operation: update_inventory {{(pid=61970) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 913.878904] env[61970]: DEBUG nova.compute.provider_tree [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 914.002028] env[61970]: DEBUG nova.network.neutron [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Successfully updated port: 9a1c8b46-0249-4e0b-94de-b0d0062ca69c {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 914.233038] env[61970]: DEBUG oslo_vmware.api [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355817, 'name': PowerOnVM_Task, 'duration_secs': 0.768892} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.233877] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 914.233877] env[61970]: INFO nova.compute.manager [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Took 8.73 seconds to spawn the instance on the hypervisor. [ 914.234188] env[61970]: DEBUG nova.compute.manager [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.234985] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f582579-f7d7-4784-8ed5-402c32f44905 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.351691] env[61970]: DEBUG nova.network.neutron [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Updating instance_info_cache with network_info: [{"id": "32711b6a-a9cc-472d-a634-e3b60711bffe", "address": "fa:16:3e:68:3d:bc", "network": {"id": "cba185ff-bfa3-4547-b7b0-638c3e79a24c", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-864852491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5c64cb9232fb413cbd7627dcf077e9ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32711b6a-a9", "ovs_interfaceid": "32711b6a-a9cc-472d-a634-e3b60711bffe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.385389] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.197s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.386191] env[61970]: DEBUG nova.compute.manager [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 914.505597] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.506240] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.506240] env[61970]: DEBUG nova.network.neutron [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 914.759947] env[61970]: INFO nova.compute.manager [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Took 24.18 seconds to build instance. [ 914.858685] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "refresh_cache-75762ba3-f129-424e-a7cb-962785dfab2a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.858685] env[61970]: DEBUG nova.objects.instance [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lazy-loading 'migration_context' on Instance uuid 75762ba3-f129-424e-a7cb-962785dfab2a {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.892785] env[61970]: DEBUG nova.compute.utils [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 914.895391] env[61970]: DEBUG nova.compute.manager [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 914.895636] env[61970]: DEBUG nova.network.neutron [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 914.972728] env[61970]: DEBUG nova.policy [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64082edf5f2e4621b5f3586e1792e42a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9235eeeb6fa4ccf959ffbea456f3694', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 915.062334] env[61970]: DEBUG nova.network.neutron [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 915.262855] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0037116d-6265-4ba1-b47f-5cff0e786326 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "7f59000a-94d9-45b6-aa7a-300d95793615" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.698s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.337101] env[61970]: DEBUG nova.network.neutron [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updating instance_info_cache with network_info: [{"id": "9a1c8b46-0249-4e0b-94de-b0d0062ca69c", "address": "fa:16:3e:93:5e:22", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a1c8b46-02", "ovs_interfaceid": "9a1c8b46-0249-4e0b-94de-b0d0062ca69c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.362510] env[61970]: DEBUG nova.objects.base [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Object Instance<75762ba3-f129-424e-a7cb-962785dfab2a> lazy-loaded attributes: info_cache,migration_context {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 915.365181] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce5782f8-e95f-4a13-add2-e3199b0852d2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.398966] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.400886] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.400886] env[61970]: DEBUG nova.network.neutron [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 915.405135] env[61970]: DEBUG nova.compute.manager [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 915.410801] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-122df865-32d6-42b7-8ce0-baf1ffbc4b39 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.424020] env[61970]: DEBUG oslo_vmware.api [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 915.424020] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52464852-2187-50da-701e-f4441bfeb0aa" [ 915.424020] env[61970]: _type = "Task" [ 915.424020] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.429267] env[61970]: DEBUG nova.network.neutron [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Successfully created port: 10059ea8-815e-48c8-9dcb-8da7857f570a {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 915.442031] env[61970]: DEBUG oslo_vmware.api [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52464852-2187-50da-701e-f4441bfeb0aa, 'name': SearchDatastore_Task, 'duration_secs': 0.009584} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.442552] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.442972] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.701815] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Acquiring lock "e4e85cff-49e7-4306-b5bc-15a798fc7e7e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.702143] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Lock "e4e85cff-49e7-4306-b5bc-15a798fc7e7e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.708099] env[61970]: DEBUG nova.compute.manager [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Received event network-changed-9a1c8b46-0249-4e0b-94de-b0d0062ca69c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 915.709205] env[61970]: DEBUG nova.compute.manager [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Refreshing instance network info cache due to event network-changed-9a1c8b46-0249-4e0b-94de-b0d0062ca69c. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 915.709205] env[61970]: DEBUG oslo_concurrency.lockutils [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] Acquiring lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.843024] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.843024] env[61970]: DEBUG nova.compute.manager [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Instance network_info: |[{"id": "9a1c8b46-0249-4e0b-94de-b0d0062ca69c", "address": "fa:16:3e:93:5e:22", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a1c8b46-02", "ovs_interfaceid": "9a1c8b46-0249-4e0b-94de-b0d0062ca69c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 915.843024] env[61970]: DEBUG oslo_concurrency.lockutils [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] Acquired lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.843024] env[61970]: DEBUG nova.network.neutron [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Refreshing network info cache for port 9a1c8b46-0249-4e0b-94de-b0d0062ca69c {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 915.843817] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:5e:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '298bb8ef-4765-494c-b157-7a349218bd1e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9a1c8b46-0249-4e0b-94de-b0d0062ca69c', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 915.852779] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Creating folder: Project (09033ee688384ca287b4fdad6e67cb1f). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 915.854150] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5984f4fa-f70f-451c-91c8-817e5eb6e743 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.868429] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Created folder: Project (09033ee688384ca287b4fdad6e67cb1f) in parent group-v288740. [ 915.868680] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Creating folder: Instances. Parent ref: group-v288828. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 915.868946] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-820958ff-53d6-47c1-a07c-7b69afb68878 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.881993] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Created folder: Instances in parent group-v288828. [ 915.882561] env[61970]: DEBUG oslo.service.loopingcall [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.882868] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 915.883151] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c371bed8-c1a5-4085-a836-454e788c119d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.917754] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 915.917754] env[61970]: value = "task-1355823" [ 915.917754] env[61970]: _type = "Task" [ 915.917754] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.933823] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355823, 'name': CreateVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.071636] env[61970]: DEBUG nova.compute.manager [req-c6fba204-2a89-4177-add8-e4bd798a3ce4 req-30413056-e3a1-4bf5-b8c6-9dea1fbe729a service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Received event network-changed-5c475b00-f4ed-4e09-916c-6f3ca7845784 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.071850] env[61970]: DEBUG nova.compute.manager [req-c6fba204-2a89-4177-add8-e4bd798a3ce4 req-30413056-e3a1-4bf5-b8c6-9dea1fbe729a service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Refreshing instance network info cache due to event network-changed-5c475b00-f4ed-4e09-916c-6f3ca7845784. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 916.073126] env[61970]: DEBUG oslo_concurrency.lockutils [req-c6fba204-2a89-4177-add8-e4bd798a3ce4 req-30413056-e3a1-4bf5-b8c6-9dea1fbe729a service nova] Acquiring lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.073126] env[61970]: DEBUG oslo_concurrency.lockutils [req-c6fba204-2a89-4177-add8-e4bd798a3ce4 req-30413056-e3a1-4bf5-b8c6-9dea1fbe729a service nova] Acquired lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.073126] env[61970]: DEBUG nova.network.neutron [req-c6fba204-2a89-4177-add8-e4bd798a3ce4 req-30413056-e3a1-4bf5-b8c6-9dea1fbe729a service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Refreshing network info cache for port 5c475b00-f4ed-4e09-916c-6f3ca7845784 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 916.160044] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.160359] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.160670] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.160998] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.161198] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.164653] env[61970]: INFO nova.compute.manager [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Terminating instance [ 916.167505] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe9c926-06ed-4bdf-838c-81538d77db97 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.172499] env[61970]: DEBUG nova.compute.manager [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 916.172733] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 916.173578] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32160f66-362a-48bd-848c-91bbee43eff3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.185369] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d37cd59-e527-49b2-9b94-73d6091fdb9e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.189143] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 916.189442] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3c8e9476-e783-44b3-af9e-7769d8ce1949 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.198627] env[61970]: DEBUG oslo_vmware.api [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 916.198627] env[61970]: value = "task-1355825" [ 916.198627] env[61970]: _type = "Task" [ 916.198627] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.235022] env[61970]: DEBUG nova.compute.manager [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 916.245084] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f61c29a-2d3e-41d6-8526-2f71d1d08cba {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.256531] env[61970]: DEBUG oslo_vmware.api [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355825, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.260705] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d1afa0-c708-4fc6-83cf-172236013aae {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.280150] env[61970]: DEBUG nova.compute.provider_tree [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.357423] env[61970]: INFO nova.network.neutron [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Port eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 916.357708] env[61970]: INFO nova.network.neutron [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Port b47e01b1-0321-4bbd-8ce3-d19568926d99 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 916.358115] env[61970]: DEBUG nova.network.neutron [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Updating instance_info_cache with network_info: [{"id": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "address": "fa:16:3e:14:8c:ec", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64a3e684-f5", "ovs_interfaceid": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.426492] env[61970]: DEBUG nova.compute.manager [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 916.433442] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355823, 'name': CreateVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.467225] env[61970]: DEBUG nova.virt.hardware [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 916.467566] env[61970]: DEBUG nova.virt.hardware [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 916.467769] env[61970]: DEBUG nova.virt.hardware [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 916.467979] env[61970]: DEBUG nova.virt.hardware [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 916.468182] env[61970]: DEBUG nova.virt.hardware [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 916.468379] env[61970]: DEBUG nova.virt.hardware [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 916.468618] env[61970]: DEBUG nova.virt.hardware [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 916.468805] env[61970]: DEBUG nova.virt.hardware [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 916.468981] env[61970]: DEBUG nova.virt.hardware [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 916.469192] env[61970]: DEBUG nova.virt.hardware [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 916.469395] env[61970]: DEBUG nova.virt.hardware [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 916.470422] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-137831ae-6634-437a-b820-9eda23292baf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.479828] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1d6433-fa0d-42e5-9603-f9ac3944d8a0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.589892] env[61970]: DEBUG nova.network.neutron [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updated VIF entry in instance network info cache for port 9a1c8b46-0249-4e0b-94de-b0d0062ca69c. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 916.590315] env[61970]: DEBUG nova.network.neutron [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updating instance_info_cache with network_info: [{"id": "9a1c8b46-0249-4e0b-94de-b0d0062ca69c", "address": "fa:16:3e:93:5e:22", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a1c8b46-02", "ovs_interfaceid": "9a1c8b46-0249-4e0b-94de-b0d0062ca69c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.762220] env[61970]: DEBUG oslo_vmware.api [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355825, 'name': PowerOffVM_Task, 'duration_secs': 0.214786} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.763298] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.763602] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 916.763767] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 916.764040] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8d438609-55da-4615-b95d-bcf7be43fc15 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.784118] env[61970]: DEBUG nova.scheduler.client.report [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.810456] env[61970]: DEBUG nova.network.neutron [req-c6fba204-2a89-4177-add8-e4bd798a3ce4 req-30413056-e3a1-4bf5-b8c6-9dea1fbe729a service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Updated VIF entry in instance network info cache for port 5c475b00-f4ed-4e09-916c-6f3ca7845784. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 916.810839] env[61970]: DEBUG nova.network.neutron [req-c6fba204-2a89-4177-add8-e4bd798a3ce4 req-30413056-e3a1-4bf5-b8c6-9dea1fbe729a service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Updating instance_info_cache with network_info: [{"id": "5c475b00-f4ed-4e09-916c-6f3ca7845784", "address": "fa:16:3e:27:f3:40", "network": {"id": "70081995-3d6c-466e-b811-1c8e532e552f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-641102750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1d4724cbb0841859369af01f2c03db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aaf1b231-3660-4453-b4f3-44d825b9a5dd", "external-id": "nsx-vlan-transportzone-6", "segmentation_id": 6, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c475b00-f4", "ovs_interfaceid": "5c475b00-f4ed-4e09-916c-6f3ca7845784", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.863582] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "refresh_cache-b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.931492] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355823, 'name': CreateVM_Task, 'duration_secs': 0.587296} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.931718] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 916.932645] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.932883] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.933340] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 916.933652] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-508946fe-5180-439c-9bd1-90ef43789448 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.939771] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 916.939771] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526c4eed-c671-5343-720b-4fea0d30a9ee" [ 916.939771] env[61970]: _type = "Task" [ 916.939771] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.949326] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526c4eed-c671-5343-720b-4fea0d30a9ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.093677] env[61970]: DEBUG oslo_concurrency.lockutils [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] Releasing lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.093677] env[61970]: DEBUG nova.compute.manager [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Received event network-vif-deleted-eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.093677] env[61970]: INFO nova.compute.manager [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Neutron deleted interface eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9; detaching it from the instance and deleting it from the info cache [ 917.093677] env[61970]: DEBUG nova.network.neutron [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Updating instance_info_cache with network_info: [{"id": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "address": "fa:16:3e:14:8c:ec", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64a3e684-f5", "ovs_interfaceid": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b47e01b1-0321-4bbd-8ce3-d19568926d99", "address": "fa:16:3e:32:95:cd", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb47e01b1-03", "ovs_interfaceid": "b47e01b1-0321-4bbd-8ce3-d19568926d99", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.153901] env[61970]: DEBUG oslo_vmware.rw_handles [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52301958-2576-ce1d-b115-9973e2019ca1/disk-0.vmdk. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 917.158836] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ead20f72-594f-4630-9ce0-829d09c50234 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.165496] env[61970]: DEBUG oslo_vmware.rw_handles [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52301958-2576-ce1d-b115-9973e2019ca1/disk-0.vmdk is in state: ready. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 917.165496] env[61970]: ERROR oslo_vmware.rw_handles [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52301958-2576-ce1d-b115-9973e2019ca1/disk-0.vmdk due to incomplete transfer. [ 917.165496] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b4c09b7b-7ab9-454d-ac91-242678e55147 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.176499] env[61970]: DEBUG oslo_vmware.rw_handles [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52301958-2576-ce1d-b115-9973e2019ca1/disk-0.vmdk. {{(pid=61970) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 917.176499] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Uploaded image c0a6b383-5802-40b2-b2d3-afa9fec0129b to the Glance image server {{(pid=61970) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 917.179516] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Destroying the VM {{(pid=61970) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 917.179833] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e1b4ce71-b474-4410-82cb-e1823f032423 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.187901] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 917.187901] env[61970]: value = "task-1355827" [ 917.187901] env[61970]: _type = "Task" [ 917.187901] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.198526] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355827, 'name': Destroy_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.313875] env[61970]: DEBUG oslo_concurrency.lockutils [req-c6fba204-2a89-4177-add8-e4bd798a3ce4 req-30413056-e3a1-4bf5-b8c6-9dea1fbe729a service nova] Releasing lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.368666] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eb628b7f-a861-43bf-a7dc-e7d56e5fd699 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "interface-b1cc1cc2-15d6-459d-9529-e592ddb225ac-eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.215s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.423969] env[61970]: DEBUG nova.network.neutron [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Successfully updated port: 10059ea8-815e-48c8-9dcb-8da7857f570a {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 917.450771] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526c4eed-c671-5343-720b-4fea0d30a9ee, 'name': SearchDatastore_Task, 'duration_secs': 0.014014} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.451034] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.451306] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 917.451563] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.451719] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.451901] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 917.452411] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-edaba4ab-c96d-4368-a06d-fe85c82acc71 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.466225] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 917.466506] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 917.467555] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a751450-355f-4f50-8f17-6cb08abb0bcb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.473984] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 917.473984] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b0691d-c33f-3ea4-8688-0865dd119fbd" [ 917.473984] env[61970]: _type = "Task" [ 917.473984] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.483271] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b0691d-c33f-3ea4-8688-0865dd119fbd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.596455] env[61970]: DEBUG oslo_concurrency.lockutils [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] Acquiring lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.700350] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355827, 'name': Destroy_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.732077] env[61970]: DEBUG nova.compute.manager [req-2a69d67e-9c08-4f0a-9093-06edca0ed617 req-3a469297-d279-49c6-9da6-fe8d18b1ce3f service nova] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Received event network-vif-plugged-10059ea8-815e-48c8-9dcb-8da7857f570a {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.732077] env[61970]: DEBUG oslo_concurrency.lockutils [req-2a69d67e-9c08-4f0a-9093-06edca0ed617 req-3a469297-d279-49c6-9da6-fe8d18b1ce3f service nova] Acquiring lock "3b346c68-53a9-4ceb-83a3-9e4bce9610b1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.732077] env[61970]: DEBUG oslo_concurrency.lockutils [req-2a69d67e-9c08-4f0a-9093-06edca0ed617 req-3a469297-d279-49c6-9da6-fe8d18b1ce3f service nova] Lock "3b346c68-53a9-4ceb-83a3-9e4bce9610b1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.732077] env[61970]: DEBUG oslo_concurrency.lockutils [req-2a69d67e-9c08-4f0a-9093-06edca0ed617 req-3a469297-d279-49c6-9da6-fe8d18b1ce3f service nova] Lock "3b346c68-53a9-4ceb-83a3-9e4bce9610b1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.732578] env[61970]: DEBUG nova.compute.manager [req-2a69d67e-9c08-4f0a-9093-06edca0ed617 req-3a469297-d279-49c6-9da6-fe8d18b1ce3f service nova] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] No waiting events found dispatching network-vif-plugged-10059ea8-815e-48c8-9dcb-8da7857f570a {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 917.732578] env[61970]: WARNING nova.compute.manager [req-2a69d67e-9c08-4f0a-9093-06edca0ed617 req-3a469297-d279-49c6-9da6-fe8d18b1ce3f service nova] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Received unexpected event network-vif-plugged-10059ea8-815e-48c8-9dcb-8da7857f570a for instance with vm_state building and task_state spawning. [ 917.732578] env[61970]: DEBUG nova.compute.manager [req-2a69d67e-9c08-4f0a-9093-06edca0ed617 req-3a469297-d279-49c6-9da6-fe8d18b1ce3f service nova] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Received event network-changed-10059ea8-815e-48c8-9dcb-8da7857f570a {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.732738] env[61970]: DEBUG nova.compute.manager [req-2a69d67e-9c08-4f0a-9093-06edca0ed617 req-3a469297-d279-49c6-9da6-fe8d18b1ce3f service nova] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Refreshing instance network info cache due to event network-changed-10059ea8-815e-48c8-9dcb-8da7857f570a. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 917.732932] env[61970]: DEBUG oslo_concurrency.lockutils [req-2a69d67e-9c08-4f0a-9093-06edca0ed617 req-3a469297-d279-49c6-9da6-fe8d18b1ce3f service nova] Acquiring lock "refresh_cache-3b346c68-53a9-4ceb-83a3-9e4bce9610b1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.733083] env[61970]: DEBUG oslo_concurrency.lockutils [req-2a69d67e-9c08-4f0a-9093-06edca0ed617 req-3a469297-d279-49c6-9da6-fe8d18b1ce3f service nova] Acquired lock "refresh_cache-3b346c68-53a9-4ceb-83a3-9e4bce9610b1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.733247] env[61970]: DEBUG nova.network.neutron [req-2a69d67e-9c08-4f0a-9093-06edca0ed617 req-3a469297-d279-49c6-9da6-fe8d18b1ce3f service nova] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Refreshing network info cache for port 10059ea8-815e-48c8-9dcb-8da7857f570a {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 917.751837] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 917.752034] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 917.752083] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Deleting the datastore file [datastore1] b1cc1cc2-15d6-459d-9529-e592ddb225ac {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 917.752764] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f398046e-a532-4059-a6a7-79303e8e7830 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.760414] env[61970]: DEBUG oslo_vmware.api [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 917.760414] env[61970]: value = "task-1355828" [ 917.760414] env[61970]: _type = "Task" [ 917.760414] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.769813] env[61970]: DEBUG oslo_vmware.api [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355828, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.795773] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.353s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.798680] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.035s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.800094] env[61970]: INFO nova.compute.claims [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 917.927179] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "refresh_cache-3b346c68-53a9-4ceb-83a3-9e4bce9610b1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.986511] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b0691d-c33f-3ea4-8688-0865dd119fbd, 'name': SearchDatastore_Task, 'duration_secs': 0.037296} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.987301] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e6e5443-1557-4bdd-82c9-8ee0fe4cbf64 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.993289] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 917.993289] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f56f27-6a8d-ffea-ad72-829604680299" [ 917.993289] env[61970]: _type = "Task" [ 917.993289] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.001735] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f56f27-6a8d-ffea-ad72-829604680299, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.199625] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355827, 'name': Destroy_Task, 'duration_secs': 0.841608} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.202019] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Destroyed the VM [ 918.202019] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Deleting Snapshot of the VM instance {{(pid=61970) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 918.202019] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c5b9fdd0-54bb-4677-bdb4-8de0beb2b5ab {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.209117] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 918.209117] env[61970]: value = "task-1355829" [ 918.209117] env[61970]: _type = "Task" [ 918.209117] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.218037] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355829, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.271195] env[61970]: DEBUG oslo_vmware.api [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355828, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.271605] env[61970]: DEBUG nova.network.neutron [req-2a69d67e-9c08-4f0a-9093-06edca0ed617 req-3a469297-d279-49c6-9da6-fe8d18b1ce3f service nova] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 918.274271] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Volume attach. Driver type: vmdk {{(pid=61970) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 918.274487] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288827', 'volume_id': '148b4432-023f-47fd-b4c0-7e51f31edf89', 'name': 'volume-148b4432-023f-47fd-b4c0-7e51f31edf89', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa', 'attached_at': '', 'detached_at': '', 'volume_id': '148b4432-023f-47fd-b4c0-7e51f31edf89', 'serial': '148b4432-023f-47fd-b4c0-7e51f31edf89'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 918.275330] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece906f4-68a0-4f32-81fe-90e71524a502 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.293591] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4d23163-71e3-4c42-b209-723638523b79 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.331920] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] volume-148b4432-023f-47fd-b4c0-7e51f31edf89/volume-148b4432-023f-47fd-b4c0-7e51f31edf89.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 918.340476] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c40e5129-b2fb-4a25-a752-c8168453c174 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.365021] env[61970]: DEBUG oslo_vmware.api [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 918.365021] env[61970]: value = "task-1355830" [ 918.365021] env[61970]: _type = "Task" [ 918.365021] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.372546] env[61970]: DEBUG oslo_vmware.api [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355830, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.417208] env[61970]: INFO nova.scheduler.client.report [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Deleted allocation for migration 1759ca51-3a57-4c63-81ae-f85ae6bc5d5f [ 918.419812] env[61970]: DEBUG nova.network.neutron [req-2a69d67e-9c08-4f0a-9093-06edca0ed617 req-3a469297-d279-49c6-9da6-fe8d18b1ce3f service nova] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.507740] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f56f27-6a8d-ffea-ad72-829604680299, 'name': SearchDatastore_Task, 'duration_secs': 0.049773} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.510306] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.510669] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] ebb92a5b-f635-4504-a57f-395de514015a/ebb92a5b-f635-4504-a57f-395de514015a.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 918.511019] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ab2dacab-221e-4abd-9b0d-9dcc0d78bee4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.518962] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 918.518962] env[61970]: value = "task-1355831" [ 918.518962] env[61970]: _type = "Task" [ 918.518962] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.529659] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355831, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.546869] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b47c38e-2168-4a06-bce4-0cd78ae2a6d8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.554776] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b183cbec-c90c-4738-a991-89dd51d7d8d7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.588407] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2707f3f0-d00e-45a1-8cc3-70f0c31afba7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.596733] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c37a41e-2bff-43e8-a422-a0752561a1ad {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.614192] env[61970]: DEBUG nova.compute.provider_tree [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.718903] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355829, 'name': RemoveSnapshot_Task, 'duration_secs': 0.415514} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.719233] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Deleted Snapshot of the VM instance {{(pid=61970) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 918.719461] env[61970]: DEBUG nova.compute.manager [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.720261] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aecb7fe-ecdf-4d1e-b248-f122f1f2b950 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.771248] env[61970]: DEBUG oslo_vmware.api [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355828, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.528098} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.771472] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 918.771816] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 918.771973] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 918.772189] env[61970]: INFO nova.compute.manager [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Took 2.60 seconds to destroy the instance on the hypervisor. [ 918.772441] env[61970]: DEBUG oslo.service.loopingcall [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.772638] env[61970]: DEBUG nova.compute.manager [-] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 918.772737] env[61970]: DEBUG nova.network.neutron [-] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 918.826646] env[61970]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port b47e01b1-0321-4bbd-8ce3-d19568926d99 could not be found.", "detail": ""}} {{(pid=61970) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 918.826882] env[61970]: DEBUG nova.network.neutron [-] Unable to show port b47e01b1-0321-4bbd-8ce3-d19568926d99 as it no longer exists. {{(pid=61970) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 918.872879] env[61970]: DEBUG oslo_vmware.api [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355830, 'name': ReconfigVM_Task, 'duration_secs': 0.36276} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.873163] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Reconfigured VM instance instance-00000048 to attach disk [datastore2] volume-148b4432-023f-47fd-b4c0-7e51f31edf89/volume-148b4432-023f-47fd-b4c0-7e51f31edf89.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 918.877880] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3988bfd2-d9a9-4a09-bead-6e7ede4e078b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.894123] env[61970]: DEBUG oslo_vmware.api [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 918.894123] env[61970]: value = "task-1355832" [ 918.894123] env[61970]: _type = "Task" [ 918.894123] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.909330] env[61970]: DEBUG oslo_vmware.api [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355832, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.923313] env[61970]: DEBUG oslo_concurrency.lockutils [req-2a69d67e-9c08-4f0a-9093-06edca0ed617 req-3a469297-d279-49c6-9da6-fe8d18b1ce3f service nova] Releasing lock "refresh_cache-3b346c68-53a9-4ceb-83a3-9e4bce9610b1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.924606] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eeb6d934-2a40-40b6-aefb-c87fff03fcdb tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "75762ba3-f129-424e-a7cb-962785dfab2a" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.553s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.925907] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "refresh_cache-3b346c68-53a9-4ceb-83a3-9e4bce9610b1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.926015] env[61970]: DEBUG nova.network.neutron [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 919.031129] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355831, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.117748] env[61970]: DEBUG nova.scheduler.client.report [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.232976] env[61970]: INFO nova.compute.manager [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Shelve offloading [ 919.235362] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 919.235756] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6cec321a-65b1-4bb6-aff8-e519871a7351 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.245675] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 919.245675] env[61970]: value = "task-1355833" [ 919.245675] env[61970]: _type = "Task" [ 919.245675] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.256858] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355833, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.410176] env[61970]: DEBUG oslo_vmware.api [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355832, 'name': ReconfigVM_Task, 'duration_secs': 0.195808} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.410176] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288827', 'volume_id': '148b4432-023f-47fd-b4c0-7e51f31edf89', 'name': 'volume-148b4432-023f-47fd-b4c0-7e51f31edf89', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa', 'attached_at': '', 'detached_at': '', 'volume_id': '148b4432-023f-47fd-b4c0-7e51f31edf89', 'serial': '148b4432-023f-47fd-b4c0-7e51f31edf89'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 919.477329] env[61970]: DEBUG nova.network.neutron [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 919.530387] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355831, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.718415} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.530688] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] ebb92a5b-f635-4504-a57f-395de514015a/ebb92a5b-f635-4504-a57f-395de514015a.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 919.530871] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 919.531132] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-55890737-7bbe-46d1-abd5-c01abd0802cb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.538833] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 919.538833] env[61970]: value = "task-1355834" [ 919.538833] env[61970]: _type = "Task" [ 919.538833] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.552737] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355834, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.624174] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.824s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.624174] env[61970]: DEBUG nova.compute.manager [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 919.659481] env[61970]: DEBUG nova.network.neutron [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Updating instance_info_cache with network_info: [{"id": "10059ea8-815e-48c8-9dcb-8da7857f570a", "address": "fa:16:3e:9c:06:2b", "network": {"id": "e02bfa6c-8f0d-4413-8dcf-e644c07792d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1410864732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9235eeeb6fa4ccf959ffbea456f3694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10059ea8-81", "ovs_interfaceid": "10059ea8-815e-48c8-9dcb-8da7857f570a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.758579] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] VM already powered off {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 919.758806] env[61970]: DEBUG nova.compute.manager [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 919.760011] env[61970]: DEBUG nova.compute.manager [req-6a1b46a5-36a2-4548-8607-ee64196466e0 req-4329617b-27d0-4029-b0f5-b6bd990e9f46 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Received event network-vif-deleted-64a3e684-f5d3-4dd2-b8c9-dc46565f71bd {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.760212] env[61970]: INFO nova.compute.manager [req-6a1b46a5-36a2-4548-8607-ee64196466e0 req-4329617b-27d0-4029-b0f5-b6bd990e9f46 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Neutron deleted interface 64a3e684-f5d3-4dd2-b8c9-dc46565f71bd; detaching it from the instance and deleting it from the info cache [ 919.760465] env[61970]: DEBUG nova.network.neutron [req-6a1b46a5-36a2-4548-8607-ee64196466e0 req-4329617b-27d0-4029-b0f5-b6bd990e9f46 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Updating instance_info_cache with network_info: [{"id": "b47e01b1-0321-4bbd-8ce3-d19568926d99", "address": "fa:16:3e:32:95:cd", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb47e01b1-03", "ovs_interfaceid": "b47e01b1-0321-4bbd-8ce3-d19568926d99", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.762212] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc05ee0-935f-4a60-a95c-7cd89a25361f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.769159] env[61970]: DEBUG oslo_concurrency.lockutils [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "refresh_cache-0c50937e-4646-47ec-ac0c-0281c9424d86" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.769340] env[61970]: DEBUG oslo_concurrency.lockutils [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "refresh_cache-0c50937e-4646-47ec-ac0c-0281c9424d86" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.769506] env[61970]: DEBUG nova.network.neutron [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 919.841736] env[61970]: DEBUG nova.network.neutron [-] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.052395] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355834, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063958} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.053240] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 920.054258] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e01126a-215f-4a0f-87ac-6b400b6d90e3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.078208] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] ebb92a5b-f635-4504-a57f-395de514015a/ebb92a5b-f635-4504-a57f-395de514015a.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 920.078877] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63c42700-a2fc-4cef-8758-814bbcadb102 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.100117] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 920.100117] env[61970]: value = "task-1355835" [ 920.100117] env[61970]: _type = "Task" [ 920.100117] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.109489] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355835, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.129569] env[61970]: DEBUG nova.compute.utils [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 920.131589] env[61970]: DEBUG nova.compute.manager [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 920.131802] env[61970]: DEBUG nova.network.neutron [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 920.162491] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "refresh_cache-3b346c68-53a9-4ceb-83a3-9e4bce9610b1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.162878] env[61970]: DEBUG nova.compute.manager [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Instance network_info: |[{"id": "10059ea8-815e-48c8-9dcb-8da7857f570a", "address": "fa:16:3e:9c:06:2b", "network": {"id": "e02bfa6c-8f0d-4413-8dcf-e644c07792d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1410864732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9235eeeb6fa4ccf959ffbea456f3694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10059ea8-81", "ovs_interfaceid": "10059ea8-815e-48c8-9dcb-8da7857f570a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 920.163432] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:06:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '510d3c47-3615-43d5-aa5d-a279fd915e71', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '10059ea8-815e-48c8-9dcb-8da7857f570a', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 920.171799] env[61970]: DEBUG oslo.service.loopingcall [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.172065] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 920.173000] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-859caae1-7ebd-4fa7-8ae5-2a1714623f1d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.188876] env[61970]: DEBUG nova.policy [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c89499bc155d4996910e814d1b3f6e23', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '88e49dea52c34ec2b4c605f7433dbf5e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 920.197354] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 920.197354] env[61970]: value = "task-1355836" [ 920.197354] env[61970]: _type = "Task" [ 920.197354] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.208847] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355836, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.266017] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6b0255ac-1a50-4b53-bc3b-449060e57a07 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.278564] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a7e278-9062-4347-87cf-1c5d66284639 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.310101] env[61970]: DEBUG nova.compute.manager [req-6a1b46a5-36a2-4548-8607-ee64196466e0 req-4329617b-27d0-4029-b0f5-b6bd990e9f46 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Detach interface failed, port_id=64a3e684-f5d3-4dd2-b8c9-dc46565f71bd, reason: Instance b1cc1cc2-15d6-459d-9529-e592ddb225ac could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 920.344260] env[61970]: INFO nova.compute.manager [-] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Took 1.57 seconds to deallocate network for instance. [ 920.448112] env[61970]: DEBUG nova.objects.instance [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lazy-loading 'flavor' on Instance uuid 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 920.539588] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "75762ba3-f129-424e-a7cb-962785dfab2a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.539903] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "75762ba3-f129-424e-a7cb-962785dfab2a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.540142] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "75762ba3-f129-424e-a7cb-962785dfab2a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.541955] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "75762ba3-f129-424e-a7cb-962785dfab2a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.541955] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "75762ba3-f129-424e-a7cb-962785dfab2a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.543348] env[61970]: INFO nova.compute.manager [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Terminating instance [ 920.549581] env[61970]: DEBUG nova.compute.manager [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 920.550419] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 920.551021] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85183e91-d49e-4a5b-9248-4d7556f34977 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.561852] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 920.562138] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c4c9b86-fe22-4f77-907a-df60bf50bc73 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.569418] env[61970]: DEBUG oslo_vmware.api [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 920.569418] env[61970]: value = "task-1355837" [ 920.569418] env[61970]: _type = "Task" [ 920.569418] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.579426] env[61970]: DEBUG oslo_vmware.api [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355837, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.610259] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355835, 'name': ReconfigVM_Task, 'duration_secs': 0.346278} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.610572] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Reconfigured VM instance instance-00000051 to attach disk [datastore2] ebb92a5b-f635-4504-a57f-395de514015a/ebb92a5b-f635-4504-a57f-395de514015a.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 920.611238] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c010217-0de6-441a-a0d7-e165797a0284 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.619867] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 920.619867] env[61970]: value = "task-1355838" [ 920.619867] env[61970]: _type = "Task" [ 920.619867] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.629456] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355838, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.632154] env[61970]: DEBUG nova.compute.manager [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 920.708620] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355836, 'name': CreateVM_Task, 'duration_secs': 0.394495} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.708810] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 920.709604] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.709793] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.710068] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 920.710336] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aacc0c61-116c-49a1-82a5-9aef94ba31fd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.717563] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 920.717563] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52db6bbb-afce-40d5-647a-b71d4be2e682" [ 920.717563] env[61970]: _type = "Task" [ 920.717563] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.719665] env[61970]: DEBUG nova.network.neutron [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Updating instance_info_cache with network_info: [{"id": "f836ffa4-f23a-4ddc-8844-4cf747f9e98d", "address": "fa:16:3e:38:a2:66", "network": {"id": "62edd125-b77c-4418-a465-1dcb3a0733c1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-371307366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6a4997482a4834a30c79c4cbab2cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf836ffa4-f2", "ovs_interfaceid": "f836ffa4-f23a-4ddc-8844-4cf747f9e98d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.733971] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52db6bbb-afce-40d5-647a-b71d4be2e682, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.758805] env[61970]: DEBUG nova.network.neutron [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Successfully created port: 56cae67e-d8f3-429f-8334-b94cb7e9046c {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 920.854767] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.854767] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.854767] env[61970]: DEBUG nova.objects.instance [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lazy-loading 'resources' on Instance uuid b1cc1cc2-15d6-459d-9529-e592ddb225ac {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 920.957009] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bebb25fc-2e78-4f6b-9de9-e2424a9b29cb tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.310s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.082255] env[61970]: DEBUG oslo_vmware.api [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355837, 'name': PowerOffVM_Task, 'duration_secs': 0.379485} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.082663] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 921.083383] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 921.083658] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d599624-649a-458a-a88d-e0010357c1b7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.129661] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355838, 'name': Rename_Task, 'duration_secs': 0.161684} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.129981] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 921.130296] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e9e96870-37c6-45e1-b904-b11e2782b3f7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.142860] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 921.142860] env[61970]: value = "task-1355840" [ 921.142860] env[61970]: _type = "Task" [ 921.142860] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.150942] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355840, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.162484] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 921.162484] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 921.162484] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Deleting the datastore file [datastore1] 75762ba3-f129-424e-a7cb-962785dfab2a {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.162904] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c59e1e85-90aa-495e-8b1d-8eda2f9a120b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.170385] env[61970]: DEBUG oslo_vmware.api [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 921.170385] env[61970]: value = "task-1355841" [ 921.170385] env[61970]: _type = "Task" [ 921.170385] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.180170] env[61970]: DEBUG oslo_vmware.api [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355841, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.184779] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "dc38f9da-9f82-4707-9764-b28aa83a8ece" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.185028] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "dc38f9da-9f82-4707-9764-b28aa83a8ece" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.226571] env[61970]: DEBUG oslo_concurrency.lockutils [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "refresh_cache-0c50937e-4646-47ec-ac0c-0281c9424d86" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.238188] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52db6bbb-afce-40d5-647a-b71d4be2e682, 'name': SearchDatastore_Task, 'duration_secs': 0.041923} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.238581] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.238870] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 921.239161] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.239329] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.239555] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 921.239874] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1af92453-580e-408b-9cc0-7524b50e43f2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.249912] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 921.250126] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 921.251839] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a27e8142-4837-4251-8daf-1760bf83f05e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.259632] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 921.259632] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b797bb-33de-1071-5481-8b7f3d0f65a1" [ 921.259632] env[61970]: _type = "Task" [ 921.259632] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.268889] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b797bb-33de-1071-5481-8b7f3d0f65a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.503091] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 921.504044] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb913e7-6662-4ba4-b54e-8934dbd1a9cb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.513262] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 921.515997] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bfe9e774-4841-4e77-ba55-b90b25847fb6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.543947] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af49778-3f5c-442d-a322-69460b6769a1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.552703] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84808534-606f-4dd0-9062-8f0a1ae8fe65 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.584157] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f76140e4-5eca-432f-a4af-27888b28f1ab {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.593162] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca00c998-f25a-4d46-b3ca-b8d73aea810d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.598782] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 921.598981] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 921.599191] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Deleting the datastore file [datastore2] 0c50937e-4646-47ec-ac0c-0281c9424d86 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.599900] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2b242ddb-8962-4b20-adbd-50ee1652928a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.609931] env[61970]: DEBUG nova.compute.provider_tree [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.613071] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 921.613071] env[61970]: value = "task-1355843" [ 921.613071] env[61970]: _type = "Task" [ 921.613071] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.623012] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355843, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.643958] env[61970]: DEBUG nova.compute.manager [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 921.658010] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355840, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.669957] env[61970]: DEBUG nova.virt.hardware [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 921.670301] env[61970]: DEBUG nova.virt.hardware [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 921.670507] env[61970]: DEBUG nova.virt.hardware [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 921.670750] env[61970]: DEBUG nova.virt.hardware [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 921.670936] env[61970]: DEBUG nova.virt.hardware [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 921.671310] env[61970]: DEBUG nova.virt.hardware [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 921.671631] env[61970]: DEBUG nova.virt.hardware [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 921.671861] env[61970]: DEBUG nova.virt.hardware [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 921.672162] env[61970]: DEBUG nova.virt.hardware [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 921.672460] env[61970]: DEBUG nova.virt.hardware [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 921.672727] env[61970]: DEBUG nova.virt.hardware [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 921.673862] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4cd3117-7b18-4791-acf0-9389262925ec {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.689777] env[61970]: DEBUG nova.compute.manager [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 921.692778] env[61970]: DEBUG oslo_vmware.api [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355841, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.288426} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.695759] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 921.695974] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 921.696174] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 921.696348] env[61970]: INFO nova.compute.manager [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Took 1.15 seconds to destroy the instance on the hypervisor. [ 921.696598] env[61970]: DEBUG oslo.service.loopingcall [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 921.696871] env[61970]: DEBUG nova.compute.manager [-] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 921.696982] env[61970]: DEBUG nova.network.neutron [-] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 921.700153] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ee77b9-fec1-4380-acb9-83494cd9656d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.773327] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b797bb-33de-1071-5481-8b7f3d0f65a1, 'name': SearchDatastore_Task, 'duration_secs': 0.01062} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.774141] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ae9b78a-182e-4965-bc6c-05899acd0503 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.781116] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 921.781116] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528fbcb4-e93d-d53b-c7bc-8eabf46769cf" [ 921.781116] env[61970]: _type = "Task" [ 921.781116] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.790269] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528fbcb4-e93d-d53b-c7bc-8eabf46769cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.066956] env[61970]: DEBUG nova.compute.manager [req-cb14af7b-6064-4e0a-8fa8-585e8de0e524 req-13af8509-9880-4c8c-8a53-b8499c6f44ad service nova] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Received event network-vif-unplugged-f836ffa4-f23a-4ddc-8844-4cf747f9e98d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 922.068424] env[61970]: DEBUG oslo_concurrency.lockutils [req-cb14af7b-6064-4e0a-8fa8-585e8de0e524 req-13af8509-9880-4c8c-8a53-b8499c6f44ad service nova] Acquiring lock "0c50937e-4646-47ec-ac0c-0281c9424d86-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.068424] env[61970]: DEBUG oslo_concurrency.lockutils [req-cb14af7b-6064-4e0a-8fa8-585e8de0e524 req-13af8509-9880-4c8c-8a53-b8499c6f44ad service nova] Lock "0c50937e-4646-47ec-ac0c-0281c9424d86-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.068424] env[61970]: DEBUG oslo_concurrency.lockutils [req-cb14af7b-6064-4e0a-8fa8-585e8de0e524 req-13af8509-9880-4c8c-8a53-b8499c6f44ad service nova] Lock "0c50937e-4646-47ec-ac0c-0281c9424d86-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.068424] env[61970]: DEBUG nova.compute.manager [req-cb14af7b-6064-4e0a-8fa8-585e8de0e524 req-13af8509-9880-4c8c-8a53-b8499c6f44ad service nova] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] No waiting events found dispatching network-vif-unplugged-f836ffa4-f23a-4ddc-8844-4cf747f9e98d {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 922.068424] env[61970]: WARNING nova.compute.manager [req-cb14af7b-6064-4e0a-8fa8-585e8de0e524 req-13af8509-9880-4c8c-8a53-b8499c6f44ad service nova] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Received unexpected event network-vif-unplugged-f836ffa4-f23a-4ddc-8844-4cf747f9e98d for instance with vm_state shelved and task_state shelving_offloading. [ 922.068948] env[61970]: DEBUG nova.compute.manager [req-cb14af7b-6064-4e0a-8fa8-585e8de0e524 req-13af8509-9880-4c8c-8a53-b8499c6f44ad service nova] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Received event network-changed-f836ffa4-f23a-4ddc-8844-4cf747f9e98d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 922.068948] env[61970]: DEBUG nova.compute.manager [req-cb14af7b-6064-4e0a-8fa8-585e8de0e524 req-13af8509-9880-4c8c-8a53-b8499c6f44ad service nova] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Refreshing instance network info cache due to event network-changed-f836ffa4-f23a-4ddc-8844-4cf747f9e98d. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 922.069182] env[61970]: DEBUG oslo_concurrency.lockutils [req-cb14af7b-6064-4e0a-8fa8-585e8de0e524 req-13af8509-9880-4c8c-8a53-b8499c6f44ad service nova] Acquiring lock "refresh_cache-0c50937e-4646-47ec-ac0c-0281c9424d86" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.069438] env[61970]: DEBUG oslo_concurrency.lockutils [req-cb14af7b-6064-4e0a-8fa8-585e8de0e524 req-13af8509-9880-4c8c-8a53-b8499c6f44ad service nova] Acquired lock "refresh_cache-0c50937e-4646-47ec-ac0c-0281c9424d86" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.069721] env[61970]: DEBUG nova.network.neutron [req-cb14af7b-6064-4e0a-8fa8-585e8de0e524 req-13af8509-9880-4c8c-8a53-b8499c6f44ad service nova] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Refreshing network info cache for port f836ffa4-f23a-4ddc-8844-4cf747f9e98d {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 922.114848] env[61970]: DEBUG nova.scheduler.client.report [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.133357] env[61970]: DEBUG oslo_vmware.api [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355843, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.341507} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.134583] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.134835] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 922.135086] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 922.154112] env[61970]: INFO nova.scheduler.client.report [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Deleted allocations for instance 0c50937e-4646-47ec-ac0c-0281c9424d86 [ 922.161575] env[61970]: DEBUG oslo_vmware.api [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355840, 'name': PowerOnVM_Task, 'duration_secs': 0.604329} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.161575] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 922.161801] env[61970]: INFO nova.compute.manager [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Took 8.95 seconds to spawn the instance on the hypervisor. [ 922.162143] env[61970]: DEBUG nova.compute.manager [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 922.162890] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c1490c-6584-4d5a-bac2-44a283605aa0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.210321] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.293219] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528fbcb4-e93d-d53b-c7bc-8eabf46769cf, 'name': SearchDatastore_Task, 'duration_secs': 0.01361} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.293547] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.293775] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 3b346c68-53a9-4ceb-83a3-9e4bce9610b1/3b346c68-53a9-4ceb-83a3-9e4bce9610b1.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 922.294062] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3eb67d72-0fc7-4cea-b36a-1c4a5733185a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.304227] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 922.304227] env[61970]: value = "task-1355844" [ 922.304227] env[61970]: _type = "Task" [ 922.304227] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.312995] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355844, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.460291] env[61970]: DEBUG nova.network.neutron [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Successfully updated port: 56cae67e-d8f3-429f-8334-b94cb7e9046c {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 922.537023] env[61970]: DEBUG nova.network.neutron [-] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.539015] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.539283] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.624790] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.771s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.626214] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.416s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.628156] env[61970]: INFO nova.compute.claims [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 922.658580] env[61970]: DEBUG oslo_concurrency.lockutils [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.659709] env[61970]: INFO nova.scheduler.client.report [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Deleted allocations for instance b1cc1cc2-15d6-459d-9529-e592ddb225ac [ 922.685329] env[61970]: INFO nova.compute.manager [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Took 15.71 seconds to build instance. [ 922.814798] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355844, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486893} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.815086] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 3b346c68-53a9-4ceb-83a3-9e4bce9610b1/3b346c68-53a9-4ceb-83a3-9e4bce9610b1.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 922.815308] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.815579] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-376757e4-afe5-45b8-ab3d-545813261770 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.821616] env[61970]: DEBUG nova.network.neutron [req-cb14af7b-6064-4e0a-8fa8-585e8de0e524 req-13af8509-9880-4c8c-8a53-b8499c6f44ad service nova] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Updated VIF entry in instance network info cache for port f836ffa4-f23a-4ddc-8844-4cf747f9e98d. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 922.821983] env[61970]: DEBUG nova.network.neutron [req-cb14af7b-6064-4e0a-8fa8-585e8de0e524 req-13af8509-9880-4c8c-8a53-b8499c6f44ad service nova] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Updating instance_info_cache with network_info: [{"id": "f836ffa4-f23a-4ddc-8844-4cf747f9e98d", "address": "fa:16:3e:38:a2:66", "network": {"id": "62edd125-b77c-4418-a465-1dcb3a0733c1", "bridge": null, "label": "tempest-DeleteServersTestJSON-371307366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6a4997482a4834a30c79c4cbab2cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapf836ffa4-f2", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.824261] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 922.824261] env[61970]: value = "task-1355845" [ 922.824261] env[61970]: _type = "Task" [ 922.824261] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.834575] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355845, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.963481] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Acquiring lock "refresh_cache-e4e85cff-49e7-4306-b5bc-15a798fc7e7e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.963685] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Acquired lock "refresh_cache-e4e85cff-49e7-4306-b5bc-15a798fc7e7e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.963905] env[61970]: DEBUG nova.network.neutron [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 923.041734] env[61970]: INFO nova.compute.manager [-] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Took 1.34 seconds to deallocate network for instance. [ 923.043112] env[61970]: DEBUG nova.compute.manager [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 923.168352] env[61970]: DEBUG oslo_concurrency.lockutils [None req-26a10503-0cc9-422a-816d-e93c04a501fb tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.007s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.169219] env[61970]: DEBUG oslo_concurrency.lockutils [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] Acquired lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.170775] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a5faac-f5da-4e6b-b586-65a61076be25 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.183616] env[61970]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 923.183755] env[61970]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=61970) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 923.184642] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e583b63b-eda7-4704-8d28-89256e31d9b2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.187467] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3fbf5b5b-113a-4607-99ea-cdf515cf1557 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "ebb92a5b-f635-4504-a57f-395de514015a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.227s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.196173] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f070412b-219c-4091-a544-9dd8a4624133 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.235488] env[61970]: ERROR root [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-288780' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-288780' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-288780' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-288780'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-288780' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-288780' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-288780'}\n"]: nova.exception.InstanceNotFound: Instance b1cc1cc2-15d6-459d-9529-e592ddb225ac could not be found. [ 923.235488] env[61970]: DEBUG oslo_concurrency.lockutils [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] Releasing lock "b1cc1cc2-15d6-459d-9529-e592ddb225ac" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.235784] env[61970]: DEBUG nova.compute.manager [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Detach interface failed, port_id=eb33aa0c-3f89-4a39-9a6c-4f0b2bcd06e9, reason: Instance b1cc1cc2-15d6-459d-9529-e592ddb225ac could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 923.236095] env[61970]: DEBUG nova.compute.manager [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Received event network-vif-deleted-b47e01b1-0321-4bbd-8ce3-d19568926d99 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.236260] env[61970]: INFO nova.compute.manager [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Neutron deleted interface b47e01b1-0321-4bbd-8ce3-d19568926d99; detaching it from the instance and deleting it from the info cache [ 923.236581] env[61970]: DEBUG nova.network.neutron [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Updating instance_info_cache with network_info: [{"id": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "address": "fa:16:3e:14:8c:ec", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64a3e684-f5", "ovs_interfaceid": "64a3e684-f5d3-4dd2-b8c9-dc46565f71bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.293741] env[61970]: DEBUG oslo_concurrency.lockutils [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "fa8bf4b0-6955-4932-a85d-5a33947ae594" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.293978] env[61970]: DEBUG oslo_concurrency.lockutils [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "fa8bf4b0-6955-4932-a85d-5a33947ae594" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.326652] env[61970]: DEBUG oslo_concurrency.lockutils [req-cb14af7b-6064-4e0a-8fa8-585e8de0e524 req-13af8509-9880-4c8c-8a53-b8499c6f44ad service nova] Releasing lock "refresh_cache-0c50937e-4646-47ec-ac0c-0281c9424d86" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.336954] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355845, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068267} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.337222] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 923.338093] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f66f93-8d3b-4f8b-935a-27c2e5586830 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.361691] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 3b346c68-53a9-4ceb-83a3-9e4bce9610b1/3b346c68-53a9-4ceb-83a3-9e4bce9610b1.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.362386] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24701050-7cf7-4318-9408-4e2e70402758 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.384427] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 923.384427] env[61970]: value = "task-1355846" [ 923.384427] env[61970]: _type = "Task" [ 923.384427] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.393441] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355846, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.501600] env[61970]: DEBUG nova.network.neutron [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 923.508089] env[61970]: DEBUG nova.compute.manager [req-67771cfd-483d-4efb-954c-73ffbbb5545b req-b32a7a9b-cba5-4cd1-80c4-981bc0b740ee service nova] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Received event network-changed-9a1c8b46-0249-4e0b-94de-b0d0062ca69c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.508089] env[61970]: DEBUG nova.compute.manager [req-67771cfd-483d-4efb-954c-73ffbbb5545b req-b32a7a9b-cba5-4cd1-80c4-981bc0b740ee service nova] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Refreshing instance network info cache due to event network-changed-9a1c8b46-0249-4e0b-94de-b0d0062ca69c. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 923.508346] env[61970]: DEBUG oslo_concurrency.lockutils [req-67771cfd-483d-4efb-954c-73ffbbb5545b req-b32a7a9b-cba5-4cd1-80c4-981bc0b740ee service nova] Acquiring lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.508346] env[61970]: DEBUG oslo_concurrency.lockutils [req-67771cfd-483d-4efb-954c-73ffbbb5545b req-b32a7a9b-cba5-4cd1-80c4-981bc0b740ee service nova] Acquired lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.508574] env[61970]: DEBUG nova.network.neutron [req-67771cfd-483d-4efb-954c-73ffbbb5545b req-b32a7a9b-cba5-4cd1-80c4-981bc0b740ee service nova] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Refreshing network info cache for port 9a1c8b46-0249-4e0b-94de-b0d0062ca69c {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 923.551759] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.569349] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.646499] env[61970]: DEBUG nova.network.neutron [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Updating instance_info_cache with network_info: [{"id": "56cae67e-d8f3-429f-8334-b94cb7e9046c", "address": "fa:16:3e:4f:dd:8a", "network": {"id": "c7535c24-a3e9-499b-81bd-52c2544ffbdb", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-343460305-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88e49dea52c34ec2b4c605f7433dbf5e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f", "external-id": "nsx-vlan-transportzone-584", "segmentation_id": 584, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56cae67e-d8", "ovs_interfaceid": "56cae67e-d8f3-429f-8334-b94cb7e9046c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.739588] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c9f36f51-e625-4175-beb5-7bebfd43ae88 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.763859] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50500ebd-3ee5-4a11-ad7a-b5b9e39bd141 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.800582] env[61970]: DEBUG nova.compute.manager [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 923.803188] env[61970]: DEBUG nova.compute.manager [req-a61d6504-8f8d-4c52-b103-36f6582ce02f req-b451edd2-2e7c-409c-b672-84d175c19d34 service nova] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Detach interface failed, port_id=b47e01b1-0321-4bbd-8ce3-d19568926d99, reason: Instance b1cc1cc2-15d6-459d-9529-e592ddb225ac could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 923.816171] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4c2260-161a-418c-b45f-db48b3cdfafb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.824203] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6852880-30d4-4db4-a52f-2023868b1ba8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.854666] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a7a06e7-d611-4589-9bcf-0fe7a04fe514 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.863556] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7adf5ace-7155-46f6-a45e-b2eede74b67f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.878906] env[61970]: DEBUG nova.compute.provider_tree [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.896609] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355846, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.991145] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d4f5a325-8eea-4a48-960f-17164a111ae5 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "0c50937e-4646-47ec-ac0c-0281c9424d86" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.096995] env[61970]: DEBUG nova.compute.manager [req-17e25326-1f97-4e88-96a9-99b595632d71 req-357f3e07-dea1-4620-b31a-5a1070367494 service nova] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Received event network-vif-deleted-32711b6a-a9cc-472d-a634-e3b60711bffe {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 924.097231] env[61970]: DEBUG nova.compute.manager [req-17e25326-1f97-4e88-96a9-99b595632d71 req-357f3e07-dea1-4620-b31a-5a1070367494 service nova] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Received event network-vif-plugged-56cae67e-d8f3-429f-8334-b94cb7e9046c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 924.097456] env[61970]: DEBUG oslo_concurrency.lockutils [req-17e25326-1f97-4e88-96a9-99b595632d71 req-357f3e07-dea1-4620-b31a-5a1070367494 service nova] Acquiring lock "e4e85cff-49e7-4306-b5bc-15a798fc7e7e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.099673] env[61970]: DEBUG oslo_concurrency.lockutils [req-17e25326-1f97-4e88-96a9-99b595632d71 req-357f3e07-dea1-4620-b31a-5a1070367494 service nova] Lock "e4e85cff-49e7-4306-b5bc-15a798fc7e7e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.099874] env[61970]: DEBUG oslo_concurrency.lockutils [req-17e25326-1f97-4e88-96a9-99b595632d71 req-357f3e07-dea1-4620-b31a-5a1070367494 service nova] Lock "e4e85cff-49e7-4306-b5bc-15a798fc7e7e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.100066] env[61970]: DEBUG nova.compute.manager [req-17e25326-1f97-4e88-96a9-99b595632d71 req-357f3e07-dea1-4620-b31a-5a1070367494 service nova] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] No waiting events found dispatching network-vif-plugged-56cae67e-d8f3-429f-8334-b94cb7e9046c {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 924.100245] env[61970]: WARNING nova.compute.manager [req-17e25326-1f97-4e88-96a9-99b595632d71 req-357f3e07-dea1-4620-b31a-5a1070367494 service nova] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Received unexpected event network-vif-plugged-56cae67e-d8f3-429f-8334-b94cb7e9046c for instance with vm_state building and task_state spawning. [ 924.100412] env[61970]: DEBUG nova.compute.manager [req-17e25326-1f97-4e88-96a9-99b595632d71 req-357f3e07-dea1-4620-b31a-5a1070367494 service nova] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Received event network-changed-56cae67e-d8f3-429f-8334-b94cb7e9046c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 924.100571] env[61970]: DEBUG nova.compute.manager [req-17e25326-1f97-4e88-96a9-99b595632d71 req-357f3e07-dea1-4620-b31a-5a1070367494 service nova] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Refreshing instance network info cache due to event network-changed-56cae67e-d8f3-429f-8334-b94cb7e9046c. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 924.100749] env[61970]: DEBUG oslo_concurrency.lockutils [req-17e25326-1f97-4e88-96a9-99b595632d71 req-357f3e07-dea1-4620-b31a-5a1070367494 service nova] Acquiring lock "refresh_cache-e4e85cff-49e7-4306-b5bc-15a798fc7e7e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.151933] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Releasing lock "refresh_cache-e4e85cff-49e7-4306-b5bc-15a798fc7e7e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.152233] env[61970]: DEBUG nova.compute.manager [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Instance network_info: |[{"id": "56cae67e-d8f3-429f-8334-b94cb7e9046c", "address": "fa:16:3e:4f:dd:8a", "network": {"id": "c7535c24-a3e9-499b-81bd-52c2544ffbdb", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-343460305-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88e49dea52c34ec2b4c605f7433dbf5e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f", "external-id": "nsx-vlan-transportzone-584", "segmentation_id": 584, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56cae67e-d8", "ovs_interfaceid": "56cae67e-d8f3-429f-8334-b94cb7e9046c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 924.152549] env[61970]: DEBUG oslo_concurrency.lockutils [req-17e25326-1f97-4e88-96a9-99b595632d71 req-357f3e07-dea1-4620-b31a-5a1070367494 service nova] Acquired lock "refresh_cache-e4e85cff-49e7-4306-b5bc-15a798fc7e7e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.152735] env[61970]: DEBUG nova.network.neutron [req-17e25326-1f97-4e88-96a9-99b595632d71 req-357f3e07-dea1-4620-b31a-5a1070367494 service nova] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Refreshing network info cache for port 56cae67e-d8f3-429f-8334-b94cb7e9046c {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 924.154040] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:dd:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '56cae67e-d8f3-429f-8334-b94cb7e9046c', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 924.161464] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Creating folder: Project (88e49dea52c34ec2b4c605f7433dbf5e). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 924.162173] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-abd87581-2718-4015-a516-9d9247da62ff {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.176379] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Created folder: Project (88e49dea52c34ec2b4c605f7433dbf5e) in parent group-v288740. [ 924.176664] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Creating folder: Instances. Parent ref: group-v288832. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 924.176863] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c5fbc5be-c94a-4b1e-8bad-393d2a4964db {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.187836] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Created folder: Instances in parent group-v288832. [ 924.188127] env[61970]: DEBUG oslo.service.loopingcall [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.188349] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 924.188625] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9584d5af-e734-4b78-afac-4bc1e508c31f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.211810] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 924.211810] env[61970]: value = "task-1355849" [ 924.211810] env[61970]: _type = "Task" [ 924.211810] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.223747] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355849, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.324212] env[61970]: DEBUG oslo_concurrency.lockutils [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.330042] env[61970]: DEBUG nova.network.neutron [req-67771cfd-483d-4efb-954c-73ffbbb5545b req-b32a7a9b-cba5-4cd1-80c4-981bc0b740ee service nova] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updated VIF entry in instance network info cache for port 9a1c8b46-0249-4e0b-94de-b0d0062ca69c. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 924.330523] env[61970]: DEBUG nova.network.neutron [req-67771cfd-483d-4efb-954c-73ffbbb5545b req-b32a7a9b-cba5-4cd1-80c4-981bc0b740ee service nova] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updating instance_info_cache with network_info: [{"id": "9a1c8b46-0249-4e0b-94de-b0d0062ca69c", "address": "fa:16:3e:93:5e:22", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a1c8b46-02", "ovs_interfaceid": "9a1c8b46-0249-4e0b-94de-b0d0062ca69c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.382875] env[61970]: DEBUG nova.scheduler.client.report [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 924.396576] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355846, 'name': ReconfigVM_Task, 'duration_secs': 0.696593} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.396886] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 3b346c68-53a9-4ceb-83a3-9e4bce9610b1/3b346c68-53a9-4ceb-83a3-9e4bce9610b1.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.398220] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-760e6972-e647-4eaa-bfb7-702091d64804 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.406648] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 924.406648] env[61970]: value = "task-1355850" [ 924.406648] env[61970]: _type = "Task" [ 924.406648] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.417414] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355850, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.722641] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355849, 'name': CreateVM_Task, 'duration_secs': 0.348832} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.722862] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 924.723562] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.723733] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.724073] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 924.724338] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd51ef1a-c8df-4408-aeb2-d1e3388c0661 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.731430] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Waiting for the task: (returnval){ [ 924.731430] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b99999-a2a3-30f6-d8e2-f8237a1a6894" [ 924.731430] env[61970]: _type = "Task" [ 924.731430] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.739873] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b99999-a2a3-30f6-d8e2-f8237a1a6894, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.835881] env[61970]: DEBUG oslo_concurrency.lockutils [req-67771cfd-483d-4efb-954c-73ffbbb5545b req-b32a7a9b-cba5-4cd1-80c4-981bc0b740ee service nova] Releasing lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.876736] env[61970]: DEBUG nova.network.neutron [req-17e25326-1f97-4e88-96a9-99b595632d71 req-357f3e07-dea1-4620-b31a-5a1070367494 service nova] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Updated VIF entry in instance network info cache for port 56cae67e-d8f3-429f-8334-b94cb7e9046c. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 924.877146] env[61970]: DEBUG nova.network.neutron [req-17e25326-1f97-4e88-96a9-99b595632d71 req-357f3e07-dea1-4620-b31a-5a1070367494 service nova] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Updating instance_info_cache with network_info: [{"id": "56cae67e-d8f3-429f-8334-b94cb7e9046c", "address": "fa:16:3e:4f:dd:8a", "network": {"id": "c7535c24-a3e9-499b-81bd-52c2544ffbdb", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-343460305-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88e49dea52c34ec2b4c605f7433dbf5e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f", "external-id": "nsx-vlan-transportzone-584", "segmentation_id": 584, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56cae67e-d8", "ovs_interfaceid": "56cae67e-d8f3-429f-8334-b94cb7e9046c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.891818] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.266s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.892380] env[61970]: DEBUG nova.compute.manager [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 924.894831] env[61970]: DEBUG oslo_concurrency.lockutils [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.236s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.895071] env[61970]: DEBUG nova.objects.instance [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lazy-loading 'resources' on Instance uuid 0c50937e-4646-47ec-ac0c-0281c9424d86 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.916931] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355850, 'name': Rename_Task, 'duration_secs': 0.158556} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.917190] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 924.917440] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-896e9635-7dba-4d23-9414-0acac6834377 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.924743] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 924.924743] env[61970]: value = "task-1355851" [ 924.924743] env[61970]: _type = "Task" [ 924.924743] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.933246] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355851, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.243192] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b99999-a2a3-30f6-d8e2-f8237a1a6894, 'name': SearchDatastore_Task, 'duration_secs': 0.012196} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.243665] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.243665] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 925.243923] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.244112] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.244319] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 925.244611] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-461e7a62-2026-4d23-99ba-70d81a47ebcb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.259970] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 925.260218] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 925.261031] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5e7adb6-0a36-4bef-8d13-f9cd53a9f767 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.266940] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Waiting for the task: (returnval){ [ 925.266940] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5214247f-c794-443b-824c-a15754848138" [ 925.266940] env[61970]: _type = "Task" [ 925.266940] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.277354] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5214247f-c794-443b-824c-a15754848138, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.380540] env[61970]: DEBUG oslo_concurrency.lockutils [req-17e25326-1f97-4e88-96a9-99b595632d71 req-357f3e07-dea1-4620-b31a-5a1070367494 service nova] Releasing lock "refresh_cache-e4e85cff-49e7-4306-b5bc-15a798fc7e7e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.398031] env[61970]: DEBUG nova.compute.utils [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 925.399330] env[61970]: DEBUG nova.compute.manager [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 925.399497] env[61970]: DEBUG nova.network.neutron [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 925.401474] env[61970]: DEBUG nova.objects.instance [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lazy-loading 'numa_topology' on Instance uuid 0c50937e-4646-47ec-ac0c-0281c9424d86 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.434813] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355851, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.446560] env[61970]: DEBUG nova.policy [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b45dc9d71c0e4693ba42322efde6c572', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5c64cb9232fb413cbd7627dcf077e9ef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 925.737708] env[61970]: DEBUG nova.network.neutron [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Successfully created port: 313663a2-7d1f-40ea-abfc-c540d35235ca {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 925.778453] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5214247f-c794-443b-824c-a15754848138, 'name': SearchDatastore_Task, 'duration_secs': 0.009958} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.781016] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2721d55a-eef5-4e7d-b953-6028df0feaf6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.786851] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.787263] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.790053] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Waiting for the task: (returnval){ [ 925.790053] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52eefbc3-f7d2-6389-29ef-5592136683f7" [ 925.790053] env[61970]: _type = "Task" [ 925.790053] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.799521] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52eefbc3-f7d2-6389-29ef-5592136683f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.903537] env[61970]: DEBUG nova.compute.manager [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 925.907086] env[61970]: DEBUG nova.objects.base [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Object Instance<0c50937e-4646-47ec-ac0c-0281c9424d86> lazy-loaded attributes: resources,numa_topology {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 925.936533] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355851, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.059074] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b550450d-cdc7-4493-8e06-ad9d1de0d24e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.067736] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f4eb28-576e-46f6-8b8b-8fc117042d69 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.099048] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08dcda7b-5584-4358-ad86-59b474be6520 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.106461] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-686e3049-e59b-44d2-899a-527ffea64bc4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.119799] env[61970]: DEBUG nova.compute.provider_tree [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.291831] env[61970]: DEBUG nova.compute.manager [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 926.309606] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52eefbc3-f7d2-6389-29ef-5592136683f7, 'name': SearchDatastore_Task, 'duration_secs': 0.010267} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.310168] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.310618] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] e4e85cff-49e7-4306-b5bc-15a798fc7e7e/e4e85cff-49e7-4306-b5bc-15a798fc7e7e.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 926.311035] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f282b8be-5084-4407-9fe6-0a4dd41360d6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.320023] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Waiting for the task: (returnval){ [ 926.320023] env[61970]: value = "task-1355852" [ 926.320023] env[61970]: _type = "Task" [ 926.320023] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.329462] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': task-1355852, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.436111] env[61970]: DEBUG oslo_vmware.api [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355851, 'name': PowerOnVM_Task, 'duration_secs': 1.018227} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.436247] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 926.436426] env[61970]: INFO nova.compute.manager [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Took 10.01 seconds to spawn the instance on the hypervisor. [ 926.436608] env[61970]: DEBUG nova.compute.manager [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.437399] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b08b766a-1fdb-4c0c-94a6-ec2188ca0d52 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.623022] env[61970]: DEBUG nova.scheduler.client.report [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 926.816393] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.830164] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': task-1355852, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.916351] env[61970]: DEBUG nova.compute.manager [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 926.937333] env[61970]: DEBUG nova.virt.hardware [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 926.937611] env[61970]: DEBUG nova.virt.hardware [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 926.937786] env[61970]: DEBUG nova.virt.hardware [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 926.937979] env[61970]: DEBUG nova.virt.hardware [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 926.938145] env[61970]: DEBUG nova.virt.hardware [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 926.938302] env[61970]: DEBUG nova.virt.hardware [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 926.938515] env[61970]: DEBUG nova.virt.hardware [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 926.938739] env[61970]: DEBUG nova.virt.hardware [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 926.938939] env[61970]: DEBUG nova.virt.hardware [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 926.939127] env[61970]: DEBUG nova.virt.hardware [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 926.939310] env[61970]: DEBUG nova.virt.hardware [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 926.940203] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f845c0fa-455b-4e17-b09f-629fb7ab4cc3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.955898] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2d88bd-53bc-4aa4-8c48-a63e5574daec {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.964236] env[61970]: INFO nova.compute.manager [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Took 17.56 seconds to build instance. [ 927.128327] env[61970]: DEBUG oslo_concurrency.lockutils [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.233s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.130791] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.579s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.130972] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.134497] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.564s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.136012] env[61970]: INFO nova.compute.claims [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 927.158053] env[61970]: INFO nova.scheduler.client.report [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Deleted allocations for instance 75762ba3-f129-424e-a7cb-962785dfab2a [ 927.335531] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': task-1355852, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.349068] env[61970]: DEBUG nova.compute.manager [req-c4d3149a-2a18-4b15-bd90-29635b7ca563 req-c4480ad8-06e6-4746-a89e-43aca5aa15bc service nova] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Received event network-vif-plugged-313663a2-7d1f-40ea-abfc-c540d35235ca {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.349230] env[61970]: DEBUG oslo_concurrency.lockutils [req-c4d3149a-2a18-4b15-bd90-29635b7ca563 req-c4480ad8-06e6-4746-a89e-43aca5aa15bc service nova] Acquiring lock "dc38f9da-9f82-4707-9764-b28aa83a8ece-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.349450] env[61970]: DEBUG oslo_concurrency.lockutils [req-c4d3149a-2a18-4b15-bd90-29635b7ca563 req-c4480ad8-06e6-4746-a89e-43aca5aa15bc service nova] Lock "dc38f9da-9f82-4707-9764-b28aa83a8ece-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.349637] env[61970]: DEBUG oslo_concurrency.lockutils [req-c4d3149a-2a18-4b15-bd90-29635b7ca563 req-c4480ad8-06e6-4746-a89e-43aca5aa15bc service nova] Lock "dc38f9da-9f82-4707-9764-b28aa83a8ece-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.349845] env[61970]: DEBUG nova.compute.manager [req-c4d3149a-2a18-4b15-bd90-29635b7ca563 req-c4480ad8-06e6-4746-a89e-43aca5aa15bc service nova] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] No waiting events found dispatching network-vif-plugged-313663a2-7d1f-40ea-abfc-c540d35235ca {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 927.350041] env[61970]: WARNING nova.compute.manager [req-c4d3149a-2a18-4b15-bd90-29635b7ca563 req-c4480ad8-06e6-4746-a89e-43aca5aa15bc service nova] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Received unexpected event network-vif-plugged-313663a2-7d1f-40ea-abfc-c540d35235ca for instance with vm_state building and task_state spawning. [ 927.454731] env[61970]: DEBUG nova.network.neutron [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Successfully updated port: 313663a2-7d1f-40ea-abfc-c540d35235ca {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 927.466228] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6acd69af-ab86-43d6-bf3b-c860bb95e4b8 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "3b346c68-53a9-4ceb-83a3-9e4bce9610b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.075s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.644954] env[61970]: DEBUG oslo_concurrency.lockutils [None req-97aeb7d5-f944-43ba-a7dc-36e4824c9697 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "0c50937e-4646-47ec-ac0c-0281c9424d86" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.627s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.645841] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d4f5a325-8eea-4a48-960f-17164a111ae5 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "0c50937e-4646-47ec-ac0c-0281c9424d86" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 3.655s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.646086] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d4f5a325-8eea-4a48-960f-17164a111ae5 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "0c50937e-4646-47ec-ac0c-0281c9424d86-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.646301] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d4f5a325-8eea-4a48-960f-17164a111ae5 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "0c50937e-4646-47ec-ac0c-0281c9424d86-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.646473] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d4f5a325-8eea-4a48-960f-17164a111ae5 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "0c50937e-4646-47ec-ac0c-0281c9424d86-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.648076] env[61970]: INFO nova.compute.manager [None req-d4f5a325-8eea-4a48-960f-17164a111ae5 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Terminating instance [ 927.649630] env[61970]: DEBUG nova.compute.manager [None req-d4f5a325-8eea-4a48-960f-17164a111ae5 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 927.649830] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f5a325-8eea-4a48-960f-17164a111ae5 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 927.650116] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d86da55e-e563-4b30-aa9d-b5c7378259b3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.660703] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f6d05bf-e5df-47d7-9a36-8f1598a44bdf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.673682] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62bb033d-a5c0-436d-961f-672e815290c3 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "75762ba3-f129-424e-a7cb-962785dfab2a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.134s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.690715] env[61970]: WARNING nova.virt.vmwareapi.vmops [None req-d4f5a325-8eea-4a48-960f-17164a111ae5 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0c50937e-4646-47ec-ac0c-0281c9424d86 could not be found. [ 927.690899] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f5a325-8eea-4a48-960f-17164a111ae5 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 927.691110] env[61970]: INFO nova.compute.manager [None req-d4f5a325-8eea-4a48-960f-17164a111ae5 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Took 0.04 seconds to destroy the instance on the hypervisor. [ 927.691388] env[61970]: DEBUG oslo.service.loopingcall [None req-d4f5a325-8eea-4a48-960f-17164a111ae5 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.692551] env[61970]: DEBUG nova.compute.manager [-] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 927.692650] env[61970]: DEBUG nova.network.neutron [-] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 927.696764] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "0d21db81-a53f-4edf-a3a6-f1949cba2499" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.696991] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "0d21db81-a53f-4edf-a3a6-f1949cba2499" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.831951] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': task-1355852, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.957752] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "refresh_cache-dc38f9da-9f82-4707-9764-b28aa83a8ece" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.957921] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "refresh_cache-dc38f9da-9f82-4707-9764-b28aa83a8ece" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.958099] env[61970]: DEBUG nova.network.neutron [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 928.200088] env[61970]: DEBUG nova.compute.manager [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 928.303228] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd51c573-fa37-497a-943a-d4147ca6a2a6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.311315] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be2f3f8c-0e45-4f80-8362-8268990913a3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.343761] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f4007d-fe9d-453f-aee1-0e2f6157c857 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.351871] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': task-1355852, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.355141] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0651e16c-67f1-4c51-bb72-684d03e9b350 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.369632] env[61970]: DEBUG nova.compute.provider_tree [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.430552] env[61970]: DEBUG nova.network.neutron [-] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.493061] env[61970]: DEBUG nova.network.neutron [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 928.698294] env[61970]: DEBUG nova.network.neutron [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Updating instance_info_cache with network_info: [{"id": "313663a2-7d1f-40ea-abfc-c540d35235ca", "address": "fa:16:3e:c7:5e:f0", "network": {"id": "cba185ff-bfa3-4547-b7b0-638c3e79a24c", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-864852491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5c64cb9232fb413cbd7627dcf077e9ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap313663a2-7d", "ovs_interfaceid": "313663a2-7d1f-40ea-abfc-c540d35235ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.721414] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.851244] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': task-1355852, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.118849} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.851804] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] e4e85cff-49e7-4306-b5bc-15a798fc7e7e/e4e85cff-49e7-4306-b5bc-15a798fc7e7e.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 928.851804] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 928.852070] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-83ba5cac-6ba3-4cbd-8cee-1151cc85c49b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.859731] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Waiting for the task: (returnval){ [ 928.859731] env[61970]: value = "task-1355853" [ 928.859731] env[61970]: _type = "Task" [ 928.859731] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.869310] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': task-1355853, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.873295] env[61970]: DEBUG nova.scheduler.client.report [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 928.933329] env[61970]: INFO nova.compute.manager [-] [instance: 0c50937e-4646-47ec-ac0c-0281c9424d86] Took 1.24 seconds to deallocate network for instance. [ 929.201578] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "refresh_cache-dc38f9da-9f82-4707-9764-b28aa83a8ece" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.201941] env[61970]: DEBUG nova.compute.manager [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Instance network_info: |[{"id": "313663a2-7d1f-40ea-abfc-c540d35235ca", "address": "fa:16:3e:c7:5e:f0", "network": {"id": "cba185ff-bfa3-4547-b7b0-638c3e79a24c", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-864852491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5c64cb9232fb413cbd7627dcf077e9ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap313663a2-7d", "ovs_interfaceid": "313663a2-7d1f-40ea-abfc-c540d35235ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 929.202407] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c7:5e:f0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '40c947c4-f471-4d48-8e43-fee54198107e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '313663a2-7d1f-40ea-abfc-c540d35235ca', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 929.209823] env[61970]: DEBUG oslo.service.loopingcall [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 929.211810] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 929.211810] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5a2f0e99-e934-483e-ae43-558da419d249 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.233623] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 929.233623] env[61970]: value = "task-1355854" [ 929.233623] env[61970]: _type = "Task" [ 929.233623] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.242519] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355854, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.371008] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': task-1355853, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068079} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.371426] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 929.372398] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffe0c9c7-2cb3-4041-99e9-5d1b945a48ff {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.378707] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.245s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.379029] env[61970]: DEBUG nova.compute.manager [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 929.393605] env[61970]: DEBUG oslo_concurrency.lockutils [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.069s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.395305] env[61970]: INFO nova.compute.claims [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 929.407799] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] e4e85cff-49e7-4306-b5bc-15a798fc7e7e/e4e85cff-49e7-4306-b5bc-15a798fc7e7e.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 929.409033] env[61970]: DEBUG nova.compute.manager [req-9b481a81-df0b-423c-ba05-c0d4f9667d2c req-d76f236d-3232-4d81-933d-f3f0d4d17f64 service nova] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Received event network-changed-313663a2-7d1f-40ea-abfc-c540d35235ca {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.409233] env[61970]: DEBUG nova.compute.manager [req-9b481a81-df0b-423c-ba05-c0d4f9667d2c req-d76f236d-3232-4d81-933d-f3f0d4d17f64 service nova] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Refreshing instance network info cache due to event network-changed-313663a2-7d1f-40ea-abfc-c540d35235ca. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 929.409612] env[61970]: DEBUG oslo_concurrency.lockutils [req-9b481a81-df0b-423c-ba05-c0d4f9667d2c req-d76f236d-3232-4d81-933d-f3f0d4d17f64 service nova] Acquiring lock "refresh_cache-dc38f9da-9f82-4707-9764-b28aa83a8ece" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.409612] env[61970]: DEBUG oslo_concurrency.lockutils [req-9b481a81-df0b-423c-ba05-c0d4f9667d2c req-d76f236d-3232-4d81-933d-f3f0d4d17f64 service nova] Acquired lock "refresh_cache-dc38f9da-9f82-4707-9764-b28aa83a8ece" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.409769] env[61970]: DEBUG nova.network.neutron [req-9b481a81-df0b-423c-ba05-c0d4f9667d2c req-d76f236d-3232-4d81-933d-f3f0d4d17f64 service nova] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Refreshing network info cache for port 313663a2-7d1f-40ea-abfc-c540d35235ca {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 929.411224] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bf366902-c93b-4386-8316-d00de79aa785 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.433483] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Waiting for the task: (returnval){ [ 929.433483] env[61970]: value = "task-1355855" [ 929.433483] env[61970]: _type = "Task" [ 929.433483] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.445199] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': task-1355855, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.743829] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355854, 'name': CreateVM_Task, 'duration_secs': 0.413456} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.744028] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 929.745028] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.745028] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.745258] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 929.745518] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a78172a-4842-4ff4-b554-d78e7d25b9ef {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.750780] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 929.750780] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52bbfa9d-d36f-7594-b719-61f10f2c53c3" [ 929.750780] env[61970]: _type = "Task" [ 929.750780] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.759177] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52bbfa9d-d36f-7594-b719-61f10f2c53c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.915038] env[61970]: DEBUG nova.compute.utils [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 929.916743] env[61970]: DEBUG nova.compute.manager [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 929.916743] env[61970]: DEBUG nova.network.neutron [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 929.943474] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': task-1355855, 'name': ReconfigVM_Task, 'duration_secs': 0.295739} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.943795] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Reconfigured VM instance instance-00000053 to attach disk [datastore1] e4e85cff-49e7-4306-b5bc-15a798fc7e7e/e4e85cff-49e7-4306-b5bc-15a798fc7e7e.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 929.944390] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3a750fff-8be8-495e-b533-aaf87f46b140 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.951083] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Waiting for the task: (returnval){ [ 929.951083] env[61970]: value = "task-1355856" [ 929.951083] env[61970]: _type = "Task" [ 929.951083] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.956667] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d4f5a325-8eea-4a48-960f-17164a111ae5 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "0c50937e-4646-47ec-ac0c-0281c9424d86" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.311s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.961175] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': task-1355856, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.985153] env[61970]: DEBUG nova.policy [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '67a4e38db9894750b16edcf7f188220f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '19dae2316adb4dc980bd972c8c40c034', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 930.245548] env[61970]: DEBUG nova.network.neutron [req-9b481a81-df0b-423c-ba05-c0d4f9667d2c req-d76f236d-3232-4d81-933d-f3f0d4d17f64 service nova] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Updated VIF entry in instance network info cache for port 313663a2-7d1f-40ea-abfc-c540d35235ca. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 930.245902] env[61970]: DEBUG nova.network.neutron [req-9b481a81-df0b-423c-ba05-c0d4f9667d2c req-d76f236d-3232-4d81-933d-f3f0d4d17f64 service nova] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Updating instance_info_cache with network_info: [{"id": "313663a2-7d1f-40ea-abfc-c540d35235ca", "address": "fa:16:3e:c7:5e:f0", "network": {"id": "cba185ff-bfa3-4547-b7b0-638c3e79a24c", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-864852491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5c64cb9232fb413cbd7627dcf077e9ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap313663a2-7d", "ovs_interfaceid": "313663a2-7d1f-40ea-abfc-c540d35235ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.263363] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52bbfa9d-d36f-7594-b719-61f10f2c53c3, 'name': SearchDatastore_Task, 'duration_secs': 0.010276} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.263665] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.263897] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 930.264146] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.264301] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.264482] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 930.264749] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d5f1f587-8184-4d6c-bb9e-899d1978c01d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.279415] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 930.279615] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 930.280497] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6380653-a636-48a7-bfd1-f051e06a6b99 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.287272] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 930.287272] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529b02c5-b450-08e4-7295-261464b30448" [ 930.287272] env[61970]: _type = "Task" [ 930.287272] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.296394] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529b02c5-b450-08e4-7295-261464b30448, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.411420] env[61970]: DEBUG nova.network.neutron [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Successfully created port: 28a89dd6-da27-49dd-a87c-a702154014c1 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 930.420373] env[61970]: DEBUG nova.compute.manager [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 930.463693] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': task-1355856, 'name': Rename_Task, 'duration_secs': 0.140032} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.466777] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 930.468375] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-81e05022-2ced-4878-a0ef-d0965e83631b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.476572] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Waiting for the task: (returnval){ [ 930.476572] env[61970]: value = "task-1355857" [ 930.476572] env[61970]: _type = "Task" [ 930.476572] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.490961] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': task-1355857, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.620926] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de2df72-e96c-43ea-a20c-5a789b9ef5aa {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.630267] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83dfddc1-6241-45dc-ba76-9937029d5cc8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.663521] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c9aba9-60ee-46a8-940b-3d62e096e20f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.672263] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a9ac00-dd77-42ba-8726-aa6929d2575d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.687011] env[61970]: DEBUG nova.compute.provider_tree [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.751111] env[61970]: DEBUG oslo_concurrency.lockutils [req-9b481a81-df0b-423c-ba05-c0d4f9667d2c req-d76f236d-3232-4d81-933d-f3f0d4d17f64 service nova] Releasing lock "refresh_cache-dc38f9da-9f82-4707-9764-b28aa83a8ece" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.798621] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529b02c5-b450-08e4-7295-261464b30448, 'name': SearchDatastore_Task, 'duration_secs': 0.011122} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.799535] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43478d83-833e-4df5-97a4-61435d45c1e2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.805768] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 930.805768] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52172bd5-addd-0185-9e9e-9903eae6837f" [ 930.805768] env[61970]: _type = "Task" [ 930.805768] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.816384] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52172bd5-addd-0185-9e9e-9903eae6837f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.987323] env[61970]: DEBUG oslo_vmware.api [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': task-1355857, 'name': PowerOnVM_Task, 'duration_secs': 0.500558} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.987600] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 930.987835] env[61970]: INFO nova.compute.manager [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Took 9.34 seconds to spawn the instance on the hypervisor. [ 930.988041] env[61970]: DEBUG nova.compute.manager [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 930.988828] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e980f8-ab14-4b7c-a20d-97bfa0439764 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.129770] env[61970]: DEBUG oslo_concurrency.lockutils [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "b9be8292-378b-4187-85dc-8d01e817faa3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.130027] env[61970]: DEBUG oslo_concurrency.lockutils [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "b9be8292-378b-4187-85dc-8d01e817faa3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.190764] env[61970]: DEBUG nova.scheduler.client.report [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 931.318962] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52172bd5-addd-0185-9e9e-9903eae6837f, 'name': SearchDatastore_Task, 'duration_secs': 0.013414} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.319411] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.319794] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] dc38f9da-9f82-4707-9764-b28aa83a8ece/dc38f9da-9f82-4707-9764-b28aa83a8ece.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 931.320197] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a65b2bd2-ff3a-44cd-bffb-dc607f8107e6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.328769] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 931.328769] env[61970]: value = "task-1355858" [ 931.328769] env[61970]: _type = "Task" [ 931.328769] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.341738] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355858, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.431163] env[61970]: DEBUG nova.compute.manager [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 931.457149] env[61970]: DEBUG nova.virt.hardware [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 931.457426] env[61970]: DEBUG nova.virt.hardware [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 931.457605] env[61970]: DEBUG nova.virt.hardware [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 931.457868] env[61970]: DEBUG nova.virt.hardware [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 931.458084] env[61970]: DEBUG nova.virt.hardware [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 931.458270] env[61970]: DEBUG nova.virt.hardware [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 931.458541] env[61970]: DEBUG nova.virt.hardware [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 931.458726] env[61970]: DEBUG nova.virt.hardware [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 931.458984] env[61970]: DEBUG nova.virt.hardware [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 931.459194] env[61970]: DEBUG nova.virt.hardware [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 931.459381] env[61970]: DEBUG nova.virt.hardware [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 931.460301] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d117f141-6bcc-4095-81f3-1eaed3ba384f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.468838] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b116249-5136-4c04-b2a2-0cb8e195fdd7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.506068] env[61970]: INFO nova.compute.manager [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Took 14.76 seconds to build instance. [ 931.632437] env[61970]: DEBUG nova.compute.manager [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 931.695700] env[61970]: DEBUG oslo_concurrency.lockutils [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.302s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.696232] env[61970]: DEBUG nova.compute.manager [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 931.698782] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.883s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.700359] env[61970]: INFO nova.compute.claims [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.839764] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355858, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.867326] env[61970]: DEBUG nova.compute.manager [req-917a91ed-f890-4e1e-a81e-6a36af87de27 req-c2761f51-afbc-479b-a00d-6fa909de7b28 service nova] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Received event network-vif-plugged-28a89dd6-da27-49dd-a87c-a702154014c1 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.867496] env[61970]: DEBUG oslo_concurrency.lockutils [req-917a91ed-f890-4e1e-a81e-6a36af87de27 req-c2761f51-afbc-479b-a00d-6fa909de7b28 service nova] Acquiring lock "ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.867726] env[61970]: DEBUG oslo_concurrency.lockutils [req-917a91ed-f890-4e1e-a81e-6a36af87de27 req-c2761f51-afbc-479b-a00d-6fa909de7b28 service nova] Lock "ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.867935] env[61970]: DEBUG oslo_concurrency.lockutils [req-917a91ed-f890-4e1e-a81e-6a36af87de27 req-c2761f51-afbc-479b-a00d-6fa909de7b28 service nova] Lock "ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.868303] env[61970]: DEBUG nova.compute.manager [req-917a91ed-f890-4e1e-a81e-6a36af87de27 req-c2761f51-afbc-479b-a00d-6fa909de7b28 service nova] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] No waiting events found dispatching network-vif-plugged-28a89dd6-da27-49dd-a87c-a702154014c1 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 931.868680] env[61970]: WARNING nova.compute.manager [req-917a91ed-f890-4e1e-a81e-6a36af87de27 req-c2761f51-afbc-479b-a00d-6fa909de7b28 service nova] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Received unexpected event network-vif-plugged-28a89dd6-da27-49dd-a87c-a702154014c1 for instance with vm_state building and task_state spawning. [ 932.008493] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0724b044-e8b6-4448-aca5-fb8b60c973c2 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Lock "e4e85cff-49e7-4306-b5bc-15a798fc7e7e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.306s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.044442] env[61970]: DEBUG nova.network.neutron [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Successfully updated port: 28a89dd6-da27-49dd-a87c-a702154014c1 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 932.155445] env[61970]: DEBUG oslo_concurrency.lockutils [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.204594] env[61970]: DEBUG nova.compute.utils [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 932.207982] env[61970]: DEBUG nova.compute.manager [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 932.208763] env[61970]: DEBUG nova.network.neutron [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 932.248096] env[61970]: DEBUG nova.policy [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '223ba28a2a204f8ba9e6d2e0f876ddd2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '28a078b8af9c469eb279be4da7459166', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 932.340634] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355858, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.517599] env[61970]: DEBUG nova.network.neutron [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Successfully created port: df21eb74-2d16-49a2-9a94-372df5b2f8a7 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 932.550676] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "refresh_cache-ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.550676] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquired lock "refresh_cache-ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.550676] env[61970]: DEBUG nova.network.neutron [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 932.708550] env[61970]: DEBUG nova.compute.manager [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 932.840330] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355858, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.885906] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75137bb3-2f36-4636-8616-52ea32b44f53 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.893985] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e26c4015-27ec-4015-a697-fe93664a33f7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.927074] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb9f74b-b57d-4674-bde8-ee0a131cea1f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.935090] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63a061a-4ba7-4bcf-a42d-c3edb47c2f66 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.948422] env[61970]: DEBUG nova.compute.provider_tree [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.101511] env[61970]: DEBUG nova.network.neutron [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 933.264818] env[61970]: DEBUG nova.network.neutron [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Updating instance_info_cache with network_info: [{"id": "28a89dd6-da27-49dd-a87c-a702154014c1", "address": "fa:16:3e:82:0a:58", "network": {"id": "42def9ae-4eed-41e5-a538-81eaced2cdd4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2047751481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19dae2316adb4dc980bd972c8c40c034", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c1b8689-a9b4-4972-beb9-6a1c8de1dc88", "external-id": "nsx-vlan-transportzone-455", "segmentation_id": 455, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28a89dd6-da", "ovs_interfaceid": "28a89dd6-da27-49dd-a87c-a702154014c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.341880] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355858, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.452343] env[61970]: DEBUG nova.scheduler.client.report [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.556919] env[61970]: DEBUG oslo_concurrency.lockutils [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Acquiring lock "e4e85cff-49e7-4306-b5bc-15a798fc7e7e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.557335] env[61970]: DEBUG oslo_concurrency.lockutils [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Lock "e4e85cff-49e7-4306-b5bc-15a798fc7e7e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.557451] env[61970]: DEBUG oslo_concurrency.lockutils [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Acquiring lock "e4e85cff-49e7-4306-b5bc-15a798fc7e7e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.557644] env[61970]: DEBUG oslo_concurrency.lockutils [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Lock "e4e85cff-49e7-4306-b5bc-15a798fc7e7e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.557817] env[61970]: DEBUG oslo_concurrency.lockutils [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Lock "e4e85cff-49e7-4306-b5bc-15a798fc7e7e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.560715] env[61970]: INFO nova.compute.manager [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Terminating instance [ 933.563762] env[61970]: DEBUG nova.compute.manager [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 933.563954] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 933.564841] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d63d8eaf-7be4-4b7f-b187-35c794d4769f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.574237] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 933.574548] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-946186ae-cd1f-483f-ab79-a235d519c773 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.583293] env[61970]: DEBUG oslo_vmware.api [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Waiting for the task: (returnval){ [ 933.583293] env[61970]: value = "task-1355859" [ 933.583293] env[61970]: _type = "Task" [ 933.583293] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.595262] env[61970]: DEBUG oslo_vmware.api [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': task-1355859, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.721399] env[61970]: DEBUG nova.compute.manager [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 933.748648] env[61970]: DEBUG nova.virt.hardware [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 933.748929] env[61970]: DEBUG nova.virt.hardware [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 933.749120] env[61970]: DEBUG nova.virt.hardware [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 933.749344] env[61970]: DEBUG nova.virt.hardware [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 933.749499] env[61970]: DEBUG nova.virt.hardware [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 933.749652] env[61970]: DEBUG nova.virt.hardware [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 933.749859] env[61970]: DEBUG nova.virt.hardware [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 933.750034] env[61970]: DEBUG nova.virt.hardware [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 933.750272] env[61970]: DEBUG nova.virt.hardware [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 933.750468] env[61970]: DEBUG nova.virt.hardware [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 933.750652] env[61970]: DEBUG nova.virt.hardware [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 933.751526] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d5b90d7-974d-465c-9242-73804b74bcd3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.761632] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d3c4a79-9640-4ab8-8fe6-142017fe337b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.767077] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Releasing lock "refresh_cache-ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.767381] env[61970]: DEBUG nova.compute.manager [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Instance network_info: |[{"id": "28a89dd6-da27-49dd-a87c-a702154014c1", "address": "fa:16:3e:82:0a:58", "network": {"id": "42def9ae-4eed-41e5-a538-81eaced2cdd4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2047751481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19dae2316adb4dc980bd972c8c40c034", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c1b8689-a9b4-4972-beb9-6a1c8de1dc88", "external-id": "nsx-vlan-transportzone-455", "segmentation_id": 455, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28a89dd6-da", "ovs_interfaceid": "28a89dd6-da27-49dd-a87c-a702154014c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 933.767916] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:0a:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9c1b8689-a9b4-4972-beb9-6a1c8de1dc88', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '28a89dd6-da27-49dd-a87c-a702154014c1', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 933.776183] env[61970]: DEBUG oslo.service.loopingcall [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.784686] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 933.785200] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-df3f2b7f-f8cd-4c69-a81f-6ceb59cb9dc3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.806811] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 933.806811] env[61970]: value = "task-1355860" [ 933.806811] env[61970]: _type = "Task" [ 933.806811] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.815726] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355860, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.843573] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355858, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.896875] env[61970]: DEBUG nova.compute.manager [req-6511dce4-c391-41cd-9e6b-6807ca7245a9 req-2cf62c01-562a-4a2f-98dd-6ad7ef10074b service nova] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Received event network-changed-28a89dd6-da27-49dd-a87c-a702154014c1 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.897343] env[61970]: DEBUG nova.compute.manager [req-6511dce4-c391-41cd-9e6b-6807ca7245a9 req-2cf62c01-562a-4a2f-98dd-6ad7ef10074b service nova] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Refreshing instance network info cache due to event network-changed-28a89dd6-da27-49dd-a87c-a702154014c1. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 933.897741] env[61970]: DEBUG oslo_concurrency.lockutils [req-6511dce4-c391-41cd-9e6b-6807ca7245a9 req-2cf62c01-562a-4a2f-98dd-6ad7ef10074b service nova] Acquiring lock "refresh_cache-ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.897826] env[61970]: DEBUG oslo_concurrency.lockutils [req-6511dce4-c391-41cd-9e6b-6807ca7245a9 req-2cf62c01-562a-4a2f-98dd-6ad7ef10074b service nova] Acquired lock "refresh_cache-ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.898181] env[61970]: DEBUG nova.network.neutron [req-6511dce4-c391-41cd-9e6b-6807ca7245a9 req-2cf62c01-562a-4a2f-98dd-6ad7ef10074b service nova] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Refreshing network info cache for port 28a89dd6-da27-49dd-a87c-a702154014c1 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 933.932737] env[61970]: DEBUG nova.compute.manager [req-8b617735-86f7-4b30-a72a-bb64b40a9302 req-84d4c297-8836-444e-aadc-1dae1dca8c00 service nova] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Received event network-vif-plugged-df21eb74-2d16-49a2-9a94-372df5b2f8a7 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.933433] env[61970]: DEBUG oslo_concurrency.lockutils [req-8b617735-86f7-4b30-a72a-bb64b40a9302 req-84d4c297-8836-444e-aadc-1dae1dca8c00 service nova] Acquiring lock "fa8bf4b0-6955-4932-a85d-5a33947ae594-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.933433] env[61970]: DEBUG oslo_concurrency.lockutils [req-8b617735-86f7-4b30-a72a-bb64b40a9302 req-84d4c297-8836-444e-aadc-1dae1dca8c00 service nova] Lock "fa8bf4b0-6955-4932-a85d-5a33947ae594-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.933682] env[61970]: DEBUG oslo_concurrency.lockutils [req-8b617735-86f7-4b30-a72a-bb64b40a9302 req-84d4c297-8836-444e-aadc-1dae1dca8c00 service nova] Lock "fa8bf4b0-6955-4932-a85d-5a33947ae594-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.933951] env[61970]: DEBUG nova.compute.manager [req-8b617735-86f7-4b30-a72a-bb64b40a9302 req-84d4c297-8836-444e-aadc-1dae1dca8c00 service nova] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] No waiting events found dispatching network-vif-plugged-df21eb74-2d16-49a2-9a94-372df5b2f8a7 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 933.934225] env[61970]: WARNING nova.compute.manager [req-8b617735-86f7-4b30-a72a-bb64b40a9302 req-84d4c297-8836-444e-aadc-1dae1dca8c00 service nova] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Received unexpected event network-vif-plugged-df21eb74-2d16-49a2-9a94-372df5b2f8a7 for instance with vm_state building and task_state spawning. [ 933.957436] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.258s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.958111] env[61970]: DEBUG nova.compute.manager [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 933.961971] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.241s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.963991] env[61970]: INFO nova.compute.claims [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 934.097147] env[61970]: DEBUG oslo_vmware.api [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': task-1355859, 'name': PowerOffVM_Task, 'duration_secs': 0.359576} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.097530] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 934.097746] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 934.098108] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-13654b77-ed45-452b-aa65-66c23c840994 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.213189] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 934.213562] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 934.213804] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Deleting the datastore file [datastore1] e4e85cff-49e7-4306-b5bc-15a798fc7e7e {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 934.214147] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f943643f-4b87-4df1-b614-bfb7a42828d9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.224204] env[61970]: DEBUG oslo_vmware.api [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Waiting for the task: (returnval){ [ 934.224204] env[61970]: value = "task-1355862" [ 934.224204] env[61970]: _type = "Task" [ 934.224204] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.234854] env[61970]: DEBUG oslo_vmware.api [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': task-1355862, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.320146] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355860, 'name': CreateVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.343695] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355858, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.870791} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.343986] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] dc38f9da-9f82-4707-9764-b28aa83a8ece/dc38f9da-9f82-4707-9764-b28aa83a8ece.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 934.344218] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 934.344525] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e4659e74-b380-49a3-93fd-54fd1fce5255 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.353127] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 934.353127] env[61970]: value = "task-1355863" [ 934.353127] env[61970]: _type = "Task" [ 934.353127] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.363235] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355863, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.472791] env[61970]: DEBUG nova.compute.utils [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 934.478047] env[61970]: DEBUG nova.compute.manager [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 934.478047] env[61970]: DEBUG nova.network.neutron [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 934.523750] env[61970]: DEBUG nova.network.neutron [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Successfully updated port: df21eb74-2d16-49a2-9a94-372df5b2f8a7 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 934.526321] env[61970]: DEBUG nova.policy [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2af915dc43c84940a937dfb8a04ea0b6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17ca05a1e6664430a02de563d98c1148', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 934.653783] env[61970]: DEBUG nova.network.neutron [req-6511dce4-c391-41cd-9e6b-6807ca7245a9 req-2cf62c01-562a-4a2f-98dd-6ad7ef10074b service nova] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Updated VIF entry in instance network info cache for port 28a89dd6-da27-49dd-a87c-a702154014c1. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 934.654199] env[61970]: DEBUG nova.network.neutron [req-6511dce4-c391-41cd-9e6b-6807ca7245a9 req-2cf62c01-562a-4a2f-98dd-6ad7ef10074b service nova] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Updating instance_info_cache with network_info: [{"id": "28a89dd6-da27-49dd-a87c-a702154014c1", "address": "fa:16:3e:82:0a:58", "network": {"id": "42def9ae-4eed-41e5-a538-81eaced2cdd4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2047751481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19dae2316adb4dc980bd972c8c40c034", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c1b8689-a9b4-4972-beb9-6a1c8de1dc88", "external-id": "nsx-vlan-transportzone-455", "segmentation_id": 455, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28a89dd6-da", "ovs_interfaceid": "28a89dd6-da27-49dd-a87c-a702154014c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.736569] env[61970]: DEBUG oslo_vmware.api [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Task: {'id': task-1355862, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176616} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.736845] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 934.737048] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 934.737236] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 934.737415] env[61970]: INFO nova.compute.manager [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Took 1.17 seconds to destroy the instance on the hypervisor. [ 934.737663] env[61970]: DEBUG oslo.service.loopingcall [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.737865] env[61970]: DEBUG nova.compute.manager [-] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 934.738052] env[61970]: DEBUG nova.network.neutron [-] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 934.819675] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355860, 'name': CreateVM_Task, 'duration_secs': 0.520779} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.819899] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 934.820588] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.820767] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.821111] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 934.821383] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0dda8bc-91b6-454c-a0b3-0efcf05559b8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.827325] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 934.827325] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]520ca228-4c76-d139-c825-c3e16f4dc2bf" [ 934.827325] env[61970]: _type = "Task" [ 934.827325] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.836266] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]520ca228-4c76-d139-c825-c3e16f4dc2bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.865462] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355863, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093768} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.865889] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 934.867091] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb9a76f-d5b4-4995-a20e-79a7417ee685 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.908322] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] dc38f9da-9f82-4707-9764-b28aa83a8ece/dc38f9da-9f82-4707-9764-b28aa83a8ece.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 934.909700] env[61970]: DEBUG nova.network.neutron [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Successfully created port: 5b506f12-7feb-41e2-8e18-abc9262ac87b {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 934.913147] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3c7f2bc-788e-4f5a-9b44-b03fbaeb6dbb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.938376] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 934.938376] env[61970]: value = "task-1355864" [ 934.938376] env[61970]: _type = "Task" [ 934.938376] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.949159] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355864, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.977395] env[61970]: DEBUG nova.compute.manager [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 935.033331] env[61970]: DEBUG oslo_concurrency.lockutils [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "refresh_cache-fa8bf4b0-6955-4932-a85d-5a33947ae594" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.033574] env[61970]: DEBUG oslo_concurrency.lockutils [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "refresh_cache-fa8bf4b0-6955-4932-a85d-5a33947ae594" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.033812] env[61970]: DEBUG nova.network.neutron [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 935.156742] env[61970]: DEBUG oslo_concurrency.lockutils [req-6511dce4-c391-41cd-9e6b-6807ca7245a9 req-2cf62c01-562a-4a2f-98dd-6ad7ef10074b service nova] Releasing lock "refresh_cache-ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.170561] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a3d6eaf-eab5-486f-aac6-d7a09dd4c2b5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.178903] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f3145de-06c8-406a-94da-0abc85c4236e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.211096] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-201748fd-2d2a-41cb-a033-2c424a46365a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.220515] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c6e262-805a-46e8-b2fb-2ca4733edcf9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.235302] env[61970]: DEBUG nova.compute.provider_tree [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 935.338894] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]520ca228-4c76-d139-c825-c3e16f4dc2bf, 'name': SearchDatastore_Task, 'duration_secs': 0.012324} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.338894] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.339195] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 935.339483] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.339638] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.340688] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 935.341079] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4884442c-1249-4ffa-96db-25b0a3528949 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.352808] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 935.353050] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 935.353892] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d3be677-f40a-4efa-9280-dc24f753317d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.359986] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 935.359986] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5235d265-1fbb-f4df-c943-db94aea9ad8e" [ 935.359986] env[61970]: _type = "Task" [ 935.359986] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.368765] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5235d265-1fbb-f4df-c943-db94aea9ad8e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.454185] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355864, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.469120] env[61970]: DEBUG nova.network.neutron [-] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.569612] env[61970]: DEBUG nova.network.neutron [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 935.738386] env[61970]: DEBUG nova.scheduler.client.report [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 935.744668] env[61970]: DEBUG nova.network.neutron [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Updating instance_info_cache with network_info: [{"id": "df21eb74-2d16-49a2-9a94-372df5b2f8a7", "address": "fa:16:3e:1c:34:29", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf21eb74-2d", "ovs_interfaceid": "df21eb74-2d16-49a2-9a94-372df5b2f8a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.872806] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5235d265-1fbb-f4df-c943-db94aea9ad8e, 'name': SearchDatastore_Task, 'duration_secs': 0.017156} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.872806] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f383064-6bde-4a76-88c0-815beb461b2e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.877060] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 935.877060] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52303fea-bc7c-6949-b88e-732bef718177" [ 935.877060] env[61970]: _type = "Task" [ 935.877060] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.885893] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52303fea-bc7c-6949-b88e-732bef718177, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.924479] env[61970]: DEBUG nova.compute.manager [req-8994a8e4-d208-4d35-abb5-309e581ccab8 req-3583b3b6-b116-4554-81ce-b056e17194fa service nova] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Received event network-changed-df21eb74-2d16-49a2-9a94-372df5b2f8a7 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.924722] env[61970]: DEBUG nova.compute.manager [req-8994a8e4-d208-4d35-abb5-309e581ccab8 req-3583b3b6-b116-4554-81ce-b056e17194fa service nova] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Refreshing instance network info cache due to event network-changed-df21eb74-2d16-49a2-9a94-372df5b2f8a7. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 935.924918] env[61970]: DEBUG oslo_concurrency.lockutils [req-8994a8e4-d208-4d35-abb5-309e581ccab8 req-3583b3b6-b116-4554-81ce-b056e17194fa service nova] Acquiring lock "refresh_cache-fa8bf4b0-6955-4932-a85d-5a33947ae594" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.949731] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355864, 'name': ReconfigVM_Task, 'duration_secs': 0.844992} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.949974] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Reconfigured VM instance instance-00000054 to attach disk [datastore1] dc38f9da-9f82-4707-9764-b28aa83a8ece/dc38f9da-9f82-4707-9764-b28aa83a8ece.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 935.950585] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-52926c03-10e1-4d5f-b370-6c78dd226124 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.957499] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 935.957499] env[61970]: value = "task-1355865" [ 935.957499] env[61970]: _type = "Task" [ 935.957499] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.966793] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355865, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.971294] env[61970]: INFO nova.compute.manager [-] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Took 1.23 seconds to deallocate network for instance. [ 935.988551] env[61970]: DEBUG nova.compute.manager [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 936.012824] env[61970]: DEBUG nova.virt.hardware [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 936.012824] env[61970]: DEBUG nova.virt.hardware [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 936.013020] env[61970]: DEBUG nova.virt.hardware [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.013122] env[61970]: DEBUG nova.virt.hardware [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 936.013280] env[61970]: DEBUG nova.virt.hardware [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.013429] env[61970]: DEBUG nova.virt.hardware [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 936.014021] env[61970]: DEBUG nova.virt.hardware [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 936.014021] env[61970]: DEBUG nova.virt.hardware [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 936.014021] env[61970]: DEBUG nova.virt.hardware [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 936.014210] env[61970]: DEBUG nova.virt.hardware [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 936.014307] env[61970]: DEBUG nova.virt.hardware [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 936.015372] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e49a9b0-c970-4c2a-b006-ad6ccca77c6c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.024435] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca312cff-6713-475d-80eb-4da95db24bd7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.243968] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.282s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.244538] env[61970]: DEBUG nova.compute.manager [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 936.248019] env[61970]: DEBUG oslo_concurrency.lockutils [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.092s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.248687] env[61970]: INFO nova.compute.claims [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 936.250867] env[61970]: DEBUG oslo_concurrency.lockutils [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "refresh_cache-fa8bf4b0-6955-4932-a85d-5a33947ae594" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.251811] env[61970]: DEBUG nova.compute.manager [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Instance network_info: |[{"id": "df21eb74-2d16-49a2-9a94-372df5b2f8a7", "address": "fa:16:3e:1c:34:29", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf21eb74-2d", "ovs_interfaceid": "df21eb74-2d16-49a2-9a94-372df5b2f8a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 936.251811] env[61970]: DEBUG oslo_concurrency.lockutils [req-8994a8e4-d208-4d35-abb5-309e581ccab8 req-3583b3b6-b116-4554-81ce-b056e17194fa service nova] Acquired lock "refresh_cache-fa8bf4b0-6955-4932-a85d-5a33947ae594" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.252090] env[61970]: DEBUG nova.network.neutron [req-8994a8e4-d208-4d35-abb5-309e581ccab8 req-3583b3b6-b116-4554-81ce-b056e17194fa service nova] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Refreshing network info cache for port df21eb74-2d16-49a2-9a94-372df5b2f8a7 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 936.252857] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1c:34:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b356db78-99c7-4464-822c-fc7e193f7878', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'df21eb74-2d16-49a2-9a94-372df5b2f8a7', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 936.260325] env[61970]: DEBUG oslo.service.loopingcall [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 936.260743] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 936.260898] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b746472d-04a6-4e38-861e-650986bd21f1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.282987] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 936.282987] env[61970]: value = "task-1355866" [ 936.282987] env[61970]: _type = "Task" [ 936.282987] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.292101] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355866, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.324336] env[61970]: DEBUG nova.compute.manager [req-b34404c0-5040-4ed1-bf8b-8b7bd704224d req-7438909b-6dc3-43ec-ae0e-48b16a168ce8 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Received event network-vif-plugged-5b506f12-7feb-41e2-8e18-abc9262ac87b {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.324650] env[61970]: DEBUG oslo_concurrency.lockutils [req-b34404c0-5040-4ed1-bf8b-8b7bd704224d req-7438909b-6dc3-43ec-ae0e-48b16a168ce8 service nova] Acquiring lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.324883] env[61970]: DEBUG oslo_concurrency.lockutils [req-b34404c0-5040-4ed1-bf8b-8b7bd704224d req-7438909b-6dc3-43ec-ae0e-48b16a168ce8 service nova] Lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.325071] env[61970]: DEBUG oslo_concurrency.lockutils [req-b34404c0-5040-4ed1-bf8b-8b7bd704224d req-7438909b-6dc3-43ec-ae0e-48b16a168ce8 service nova] Lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.325288] env[61970]: DEBUG nova.compute.manager [req-b34404c0-5040-4ed1-bf8b-8b7bd704224d req-7438909b-6dc3-43ec-ae0e-48b16a168ce8 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] No waiting events found dispatching network-vif-plugged-5b506f12-7feb-41e2-8e18-abc9262ac87b {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 936.325580] env[61970]: WARNING nova.compute.manager [req-b34404c0-5040-4ed1-bf8b-8b7bd704224d req-7438909b-6dc3-43ec-ae0e-48b16a168ce8 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Received unexpected event network-vif-plugged-5b506f12-7feb-41e2-8e18-abc9262ac87b for instance with vm_state building and task_state spawning. [ 936.387847] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52303fea-bc7c-6949-b88e-732bef718177, 'name': SearchDatastore_Task, 'duration_secs': 0.010782} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.388096] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.388388] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc/ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 936.388673] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ed1e8f17-79c2-45c0-8de4-2372d714089b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.396436] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 936.396436] env[61970]: value = "task-1355867" [ 936.396436] env[61970]: _type = "Task" [ 936.396436] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.404925] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355867, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.469779] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355865, 'name': Rename_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.477990] env[61970]: DEBUG oslo_concurrency.lockutils [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.753716] env[61970]: DEBUG nova.compute.utils [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 936.758428] env[61970]: DEBUG nova.compute.manager [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 936.758728] env[61970]: DEBUG nova.network.neutron [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 936.795355] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355866, 'name': CreateVM_Task, 'duration_secs': 0.380108} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.795540] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 936.796296] env[61970]: DEBUG oslo_concurrency.lockutils [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.796500] env[61970]: DEBUG oslo_concurrency.lockutils [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.796849] env[61970]: DEBUG oslo_concurrency.lockutils [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 936.797163] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d14c9943-6f01-4217-86ae-43cead985157 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.803842] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 936.803842] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5270ce9b-f191-d5dd-ae29-119beec3fd04" [ 936.803842] env[61970]: _type = "Task" [ 936.803842] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.814547] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5270ce9b-f191-d5dd-ae29-119beec3fd04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.834194] env[61970]: DEBUG nova.policy [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64082edf5f2e4621b5f3586e1792e42a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9235eeeb6fa4ccf959ffbea456f3694', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 936.909020] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355867, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470557} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.909020] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc/ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 936.909249] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 936.909515] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8a5815fc-d763-4b19-a7b3-8b79f654f787 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.919031] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 936.919031] env[61970]: value = "task-1355868" [ 936.919031] env[61970]: _type = "Task" [ 936.919031] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.927016] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355868, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.981560] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355865, 'name': Rename_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.087590] env[61970]: DEBUG nova.network.neutron [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Successfully updated port: 5b506f12-7feb-41e2-8e18-abc9262ac87b {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 937.117014] env[61970]: DEBUG nova.compute.manager [req-4e013646-cb49-47d7-9b06-67e3ef7fab62 req-df30321a-0e93-4f0e-944b-33a48e3903c2 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Received event network-changed-5b506f12-7feb-41e2-8e18-abc9262ac87b {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.117236] env[61970]: DEBUG nova.compute.manager [req-4e013646-cb49-47d7-9b06-67e3ef7fab62 req-df30321a-0e93-4f0e-944b-33a48e3903c2 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Refreshing instance network info cache due to event network-changed-5b506f12-7feb-41e2-8e18-abc9262ac87b. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 937.117455] env[61970]: DEBUG oslo_concurrency.lockutils [req-4e013646-cb49-47d7-9b06-67e3ef7fab62 req-df30321a-0e93-4f0e-944b-33a48e3903c2 service nova] Acquiring lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.117600] env[61970]: DEBUG oslo_concurrency.lockutils [req-4e013646-cb49-47d7-9b06-67e3ef7fab62 req-df30321a-0e93-4f0e-944b-33a48e3903c2 service nova] Acquired lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.117763] env[61970]: DEBUG nova.network.neutron [req-4e013646-cb49-47d7-9b06-67e3ef7fab62 req-df30321a-0e93-4f0e-944b-33a48e3903c2 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Refreshing network info cache for port 5b506f12-7feb-41e2-8e18-abc9262ac87b {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 937.176968] env[61970]: DEBUG nova.network.neutron [req-8994a8e4-d208-4d35-abb5-309e581ccab8 req-3583b3b6-b116-4554-81ce-b056e17194fa service nova] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Updated VIF entry in instance network info cache for port df21eb74-2d16-49a2-9a94-372df5b2f8a7. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 937.179353] env[61970]: DEBUG nova.network.neutron [req-8994a8e4-d208-4d35-abb5-309e581ccab8 req-3583b3b6-b116-4554-81ce-b056e17194fa service nova] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Updating instance_info_cache with network_info: [{"id": "df21eb74-2d16-49a2-9a94-372df5b2f8a7", "address": "fa:16:3e:1c:34:29", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf21eb74-2d", "ovs_interfaceid": "df21eb74-2d16-49a2-9a94-372df5b2f8a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.259471] env[61970]: DEBUG nova.compute.manager [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 937.275613] env[61970]: DEBUG nova.network.neutron [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Successfully created port: fa8c21c5-9351-49d0-8080-e693c96d1348 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 937.317466] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5270ce9b-f191-d5dd-ae29-119beec3fd04, 'name': SearchDatastore_Task, 'duration_secs': 0.048522} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.317789] env[61970]: DEBUG oslo_concurrency.lockutils [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.318078] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 937.318313] env[61970]: DEBUG oslo_concurrency.lockutils [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.318467] env[61970]: DEBUG oslo_concurrency.lockutils [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.318646] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 937.318917] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-951b8bfd-14d5-4b50-a6af-050e8456cb9b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.332142] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 937.332371] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 937.335514] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-216664ec-815b-44ee-a64a-c5d4260f3c07 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.342115] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 937.342115] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52022eea-a8ae-3323-1443-55118cceb3c7" [ 937.342115] env[61970]: _type = "Task" [ 937.342115] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.352498] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52022eea-a8ae-3323-1443-55118cceb3c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.428531] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355868, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071877} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.428794] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 937.429580] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89970da1-2420-4510-975d-3cfa483b9337 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.433230] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-277be4e6-d07d-4905-a960-571e9cd7b03d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.455185] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc/ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 937.457353] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e5ba22c-a07e-4add-912d-4185796814ea {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.473675] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae5e8a34-4cf7-4967-8f09-d2bbfc55a4ba {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.485526] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355865, 'name': Rename_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.510280] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 937.510280] env[61970]: value = "task-1355869" [ 937.510280] env[61970]: _type = "Task" [ 937.510280] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.511234] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e6c146-5f9b-4d95-87cc-75cba9826862 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.524576] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f05c7eb3-2ea0-494c-bac6-85b8a2e6730f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.528240] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355869, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.538296] env[61970]: DEBUG nova.compute.provider_tree [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.591144] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.649175] env[61970]: DEBUG nova.network.neutron [req-4e013646-cb49-47d7-9b06-67e3ef7fab62 req-df30321a-0e93-4f0e-944b-33a48e3903c2 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 937.679678] env[61970]: DEBUG oslo_concurrency.lockutils [req-8994a8e4-d208-4d35-abb5-309e581ccab8 req-3583b3b6-b116-4554-81ce-b056e17194fa service nova] Releasing lock "refresh_cache-fa8bf4b0-6955-4932-a85d-5a33947ae594" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.680013] env[61970]: DEBUG nova.compute.manager [req-8994a8e4-d208-4d35-abb5-309e581ccab8 req-3583b3b6-b116-4554-81ce-b056e17194fa service nova] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Received event network-vif-deleted-56cae67e-d8f3-429f-8334-b94cb7e9046c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.717087] env[61970]: DEBUG nova.network.neutron [req-4e013646-cb49-47d7-9b06-67e3ef7fab62 req-df30321a-0e93-4f0e-944b-33a48e3903c2 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.854535] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52022eea-a8ae-3323-1443-55118cceb3c7, 'name': SearchDatastore_Task, 'duration_secs': 0.009302} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.855423] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47f2ce0e-3248-47f8-aba9-489ae80cca85 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.861743] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 937.861743] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5289b936-3af1-8782-d2cf-7d91eb1e38ef" [ 937.861743] env[61970]: _type = "Task" [ 937.861743] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.870219] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5289b936-3af1-8782-d2cf-7d91eb1e38ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.983331] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355865, 'name': Rename_Task, 'duration_secs': 1.964405} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.983632] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 937.983883] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3b12837d-29f1-4ecf-8c4f-33b4230e8269 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.991063] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 937.991063] env[61970]: value = "task-1355870" [ 937.991063] env[61970]: _type = "Task" [ 937.991063] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.999010] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355870, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.021426] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355869, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.041690] env[61970]: DEBUG nova.scheduler.client.report [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 938.216481] env[61970]: DEBUG oslo_concurrency.lockutils [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.216741] env[61970]: DEBUG oslo_concurrency.lockutils [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.219349] env[61970]: DEBUG oslo_concurrency.lockutils [req-4e013646-cb49-47d7-9b06-67e3ef7fab62 req-df30321a-0e93-4f0e-944b-33a48e3903c2 service nova] Releasing lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.219674] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.219837] env[61970]: DEBUG nova.network.neutron [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 938.269914] env[61970]: DEBUG nova.compute.manager [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 938.295944] env[61970]: DEBUG nova.virt.hardware [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 938.296307] env[61970]: DEBUG nova.virt.hardware [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 938.296476] env[61970]: DEBUG nova.virt.hardware [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 938.296708] env[61970]: DEBUG nova.virt.hardware [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 938.296932] env[61970]: DEBUG nova.virt.hardware [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 938.297156] env[61970]: DEBUG nova.virt.hardware [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 938.297439] env[61970]: DEBUG nova.virt.hardware [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 938.297618] env[61970]: DEBUG nova.virt.hardware [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 938.297823] env[61970]: DEBUG nova.virt.hardware [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 938.298074] env[61970]: DEBUG nova.virt.hardware [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 938.298377] env[61970]: DEBUG nova.virt.hardware [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 938.299650] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f338f08b-8bdc-4210-94e4-6f4a5b65979a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.308818] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684cf2ba-13f1-4fd5-ac2a-f9c33d4217b8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.373245] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5289b936-3af1-8782-d2cf-7d91eb1e38ef, 'name': SearchDatastore_Task, 'duration_secs': 0.010108} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.373511] env[61970]: DEBUG oslo_concurrency.lockutils [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.373899] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] fa8bf4b0-6955-4932-a85d-5a33947ae594/fa8bf4b0-6955-4932-a85d-5a33947ae594.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 938.374040] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9317e56-f02d-4a17-938b-817171654d21 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.381714] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 938.381714] env[61970]: value = "task-1355871" [ 938.381714] env[61970]: _type = "Task" [ 938.381714] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.390538] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355871, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.501355] env[61970]: DEBUG oslo_vmware.api [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355870, 'name': PowerOnVM_Task, 'duration_secs': 0.442707} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.501613] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 938.501818] env[61970]: INFO nova.compute.manager [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Took 11.59 seconds to spawn the instance on the hypervisor. [ 938.502360] env[61970]: DEBUG nova.compute.manager [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 938.502955] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854c2a1d-09a8-4e01-b74f-18279c641fda {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.519872] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355869, 'name': ReconfigVM_Task, 'duration_secs': 0.656477} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.521030] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Reconfigured VM instance instance-00000055 to attach disk [datastore2] ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc/ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 938.521030] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0ac88f54-990c-47f7-88dc-c37664457373 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.528352] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 938.528352] env[61970]: value = "task-1355872" [ 938.528352] env[61970]: _type = "Task" [ 938.528352] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.537087] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355872, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.546288] env[61970]: DEBUG oslo_concurrency.lockutils [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.299s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.546794] env[61970]: DEBUG nova.compute.manager [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 938.549353] env[61970]: DEBUG oslo_concurrency.lockutils [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.071s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.549584] env[61970]: DEBUG nova.objects.instance [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Lazy-loading 'resources' on Instance uuid e4e85cff-49e7-4306-b5bc-15a798fc7e7e {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.720022] env[61970]: DEBUG nova.compute.utils [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 938.764573] env[61970]: DEBUG nova.network.neutron [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 938.835981] env[61970]: DEBUG nova.compute.manager [req-4b8f60e8-b30c-4467-8ebc-2391049b4ad2 req-d9e45c0b-3fe2-46a6-b84e-119f78038b48 service nova] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Received event network-vif-plugged-fa8c21c5-9351-49d0-8080-e693c96d1348 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.836343] env[61970]: DEBUG oslo_concurrency.lockutils [req-4b8f60e8-b30c-4467-8ebc-2391049b4ad2 req-d9e45c0b-3fe2-46a6-b84e-119f78038b48 service nova] Acquiring lock "0d21db81-a53f-4edf-a3a6-f1949cba2499-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.836530] env[61970]: DEBUG oslo_concurrency.lockutils [req-4b8f60e8-b30c-4467-8ebc-2391049b4ad2 req-d9e45c0b-3fe2-46a6-b84e-119f78038b48 service nova] Lock "0d21db81-a53f-4edf-a3a6-f1949cba2499-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.836717] env[61970]: DEBUG oslo_concurrency.lockutils [req-4b8f60e8-b30c-4467-8ebc-2391049b4ad2 req-d9e45c0b-3fe2-46a6-b84e-119f78038b48 service nova] Lock "0d21db81-a53f-4edf-a3a6-f1949cba2499-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.836898] env[61970]: DEBUG nova.compute.manager [req-4b8f60e8-b30c-4467-8ebc-2391049b4ad2 req-d9e45c0b-3fe2-46a6-b84e-119f78038b48 service nova] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] No waiting events found dispatching network-vif-plugged-fa8c21c5-9351-49d0-8080-e693c96d1348 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 938.837473] env[61970]: WARNING nova.compute.manager [req-4b8f60e8-b30c-4467-8ebc-2391049b4ad2 req-d9e45c0b-3fe2-46a6-b84e-119f78038b48 service nova] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Received unexpected event network-vif-plugged-fa8c21c5-9351-49d0-8080-e693c96d1348 for instance with vm_state building and task_state spawning. [ 938.894048] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355871, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.927683] env[61970]: DEBUG nova.network.neutron [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Successfully updated port: fa8c21c5-9351-49d0-8080-e693c96d1348 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 938.976726] env[61970]: DEBUG nova.network.neutron [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Updating instance_info_cache with network_info: [{"id": "5b506f12-7feb-41e2-8e18-abc9262ac87b", "address": "fa:16:3e:f9:43:05", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b506f12-7f", "ovs_interfaceid": "5b506f12-7feb-41e2-8e18-abc9262ac87b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.025150] env[61970]: INFO nova.compute.manager [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Took 16.83 seconds to build instance. [ 939.038727] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355872, 'name': Rename_Task, 'duration_secs': 0.235604} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.039476] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 939.039736] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-93b1aae7-5d9c-47b7-a9a2-ea2cd59181f6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.046698] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 939.046698] env[61970]: value = "task-1355873" [ 939.046698] env[61970]: _type = "Task" [ 939.046698] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.052147] env[61970]: DEBUG nova.compute.utils [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 939.058666] env[61970]: DEBUG nova.compute.manager [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 939.058836] env[61970]: DEBUG nova.network.neutron [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 939.060566] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355873, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.099946] env[61970]: DEBUG nova.policy [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd04344afddb14418b5018b69fbd7694e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c6a4997482a4834a30c79c4cbab2cc1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 939.225193] env[61970]: DEBUG oslo_concurrency.lockutils [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.255714] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d7043e-40b7-4de6-948d-73cc5f339878 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.266214] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a6f46b2-b6bc-497b-abc1-da026a2263d5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.300377] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6428296d-03cf-485b-a43e-cc6fd55290f9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.308479] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ba904b-81ce-4c21-abd3-e71b3a31d0a5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.325951] env[61970]: DEBUG nova.compute.provider_tree [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.364601] env[61970]: DEBUG nova.network.neutron [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Successfully created port: 490f6a87-12ad-4d7f-9dc5-46d694adf506 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 939.393588] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355871, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.576411} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.393849] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] fa8bf4b0-6955-4932-a85d-5a33947ae594/fa8bf4b0-6955-4932-a85d-5a33947ae594.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 939.394079] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 939.394342] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-081f4090-309f-41d4-9576-746cf78fd03e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.401561] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 939.401561] env[61970]: value = "task-1355874" [ 939.401561] env[61970]: _type = "Task" [ 939.401561] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.411426] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355874, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.430554] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "refresh_cache-0d21db81-a53f-4edf-a3a6-f1949cba2499" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.430709] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "refresh_cache-0d21db81-a53f-4edf-a3a6-f1949cba2499" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.430869] env[61970]: DEBUG nova.network.neutron [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 939.479657] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.479910] env[61970]: DEBUG nova.compute.manager [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Instance network_info: |[{"id": "5b506f12-7feb-41e2-8e18-abc9262ac87b", "address": "fa:16:3e:f9:43:05", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b506f12-7f", "ovs_interfaceid": "5b506f12-7feb-41e2-8e18-abc9262ac87b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 939.480363] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:43:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '68add7d6-c025-46fa-84d3-9c589adb63e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5b506f12-7feb-41e2-8e18-abc9262ac87b', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 939.488315] env[61970]: DEBUG oslo.service.loopingcall [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.488558] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 939.488792] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7d9f97f7-0a45-40d2-bf6c-65443ff816a7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.509892] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 939.509892] env[61970]: value = "task-1355875" [ 939.509892] env[61970]: _type = "Task" [ 939.509892] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.518249] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355875, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.526867] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2984b625-4230-4fd6-898d-0919e46710ad tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "dc38f9da-9f82-4707-9764-b28aa83a8ece" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.342s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.556555] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355873, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.562303] env[61970]: DEBUG nova.compute.manager [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 939.826752] env[61970]: DEBUG nova.scheduler.client.report [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.914067] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355874, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078671} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.914476] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 939.915317] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb64711c-5c6a-4797-b6e2-3ca5d97a612c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.939771] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] fa8bf4b0-6955-4932-a85d-5a33947ae594/fa8bf4b0-6955-4932-a85d-5a33947ae594.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.941790] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c527480-29fa-42f0-ae27-1a99fe5f7931 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.967142] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 939.967142] env[61970]: value = "task-1355876" [ 939.967142] env[61970]: _type = "Task" [ 939.967142] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.975957] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355876, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.988106] env[61970]: DEBUG nova.network.neutron [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 940.020136] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355875, 'name': CreateVM_Task} progress is 25%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.058647] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355873, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.288043] env[61970]: DEBUG oslo_concurrency.lockutils [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.288043] env[61970]: DEBUG oslo_concurrency.lockutils [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.288043] env[61970]: INFO nova.compute.manager [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Attaching volume 3916ae0c-f673-4a6a-bd2a-9134a92569c0 to /dev/sdb [ 940.326705] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61b2506f-7605-4df3-ac14-74f943a0cd50 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.331842] env[61970]: DEBUG oslo_concurrency.lockutils [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.782s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.336063] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-354bd416-3842-4ef2-9d82-11a480a4aeb3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.351346] env[61970]: DEBUG nova.virt.block_device [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Updating existing volume attachment record: 096fcdc8-858d-46b9-8262-c3a094b6b9dd {{(pid=61970) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 940.356170] env[61970]: INFO nova.scheduler.client.report [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Deleted allocations for instance e4e85cff-49e7-4306-b5bc-15a798fc7e7e [ 940.410664] env[61970]: DEBUG nova.network.neutron [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Updating instance_info_cache with network_info: [{"id": "fa8c21c5-9351-49d0-8080-e693c96d1348", "address": "fa:16:3e:ce:cf:ec", "network": {"id": "e02bfa6c-8f0d-4413-8dcf-e644c07792d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1410864732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9235eeeb6fa4ccf959ffbea456f3694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa8c21c5-93", "ovs_interfaceid": "fa8c21c5-9351-49d0-8080-e693c96d1348", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.479161] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355876, 'name': ReconfigVM_Task, 'duration_secs': 0.283471} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.479421] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Reconfigured VM instance instance-00000056 to attach disk [datastore2] fa8bf4b0-6955-4932-a85d-5a33947ae594/fa8bf4b0-6955-4932-a85d-5a33947ae594.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 940.480133] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4ad28e76-1b35-4741-a124-0f0b57a3125f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.487905] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 940.487905] env[61970]: value = "task-1355877" [ 940.487905] env[61970]: _type = "Task" [ 940.487905] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.497255] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355877, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.520852] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355875, 'name': CreateVM_Task, 'duration_secs': 0.620039} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.521070] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 940.521809] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.522012] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.522364] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 940.522962] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10858dda-69a3-4f2c-bc3a-a6640427f73e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.528025] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 940.528025] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523b7dad-05b7-cdec-54e2-50b37544a256" [ 940.528025] env[61970]: _type = "Task" [ 940.528025] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.535656] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523b7dad-05b7-cdec-54e2-50b37544a256, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.557568] env[61970]: DEBUG oslo_vmware.api [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355873, 'name': PowerOnVM_Task, 'duration_secs': 1.082918} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.557893] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 940.558059] env[61970]: INFO nova.compute.manager [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Took 9.13 seconds to spawn the instance on the hypervisor. [ 940.558287] env[61970]: DEBUG nova.compute.manager [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 940.559165] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70475a1b-9996-4dc0-8e62-51ee32a9d1a7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.572124] env[61970]: DEBUG nova.compute.manager [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 940.602335] env[61970]: DEBUG nova.virt.hardware [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 940.602335] env[61970]: DEBUG nova.virt.hardware [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 940.602335] env[61970]: DEBUG nova.virt.hardware [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.602335] env[61970]: DEBUG nova.virt.hardware [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 940.602335] env[61970]: DEBUG nova.virt.hardware [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.602335] env[61970]: DEBUG nova.virt.hardware [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 940.602335] env[61970]: DEBUG nova.virt.hardware [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 940.602639] env[61970]: DEBUG nova.virt.hardware [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 940.602639] env[61970]: DEBUG nova.virt.hardware [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 940.602743] env[61970]: DEBUG nova.virt.hardware [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 940.602899] env[61970]: DEBUG nova.virt.hardware [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 940.603854] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f374676-0ad6-4faf-b9a4-a3235c98e29d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.614725] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d96363c5-a638-49ec-8d64-f4697d6cd11c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.863264] env[61970]: DEBUG nova.compute.manager [req-0a9d5283-0429-4a00-9c55-d6c5f7137926 req-8391530e-aa55-4cd5-9107-660d498ddc87 service nova] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Received event network-changed-fa8c21c5-9351-49d0-8080-e693c96d1348 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.863264] env[61970]: DEBUG nova.compute.manager [req-0a9d5283-0429-4a00-9c55-d6c5f7137926 req-8391530e-aa55-4cd5-9107-660d498ddc87 service nova] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Refreshing instance network info cache due to event network-changed-fa8c21c5-9351-49d0-8080-e693c96d1348. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 940.863264] env[61970]: DEBUG oslo_concurrency.lockutils [req-0a9d5283-0429-4a00-9c55-d6c5f7137926 req-8391530e-aa55-4cd5-9107-660d498ddc87 service nova] Acquiring lock "refresh_cache-0d21db81-a53f-4edf-a3a6-f1949cba2499" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.866462] env[61970]: DEBUG oslo_concurrency.lockutils [None req-83f68cfb-77a7-4770-a8f4-f51899128744 tempest-ServerMetadataTestJSON-1696791573 tempest-ServerMetadataTestJSON-1696791573-project-member] Lock "e4e85cff-49e7-4306-b5bc-15a798fc7e7e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.309s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.898744] env[61970]: DEBUG nova.compute.manager [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Stashing vm_state: active {{(pid=61970) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 940.913391] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "refresh_cache-0d21db81-a53f-4edf-a3a6-f1949cba2499" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.915171] env[61970]: DEBUG nova.compute.manager [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Instance network_info: |[{"id": "fa8c21c5-9351-49d0-8080-e693c96d1348", "address": "fa:16:3e:ce:cf:ec", "network": {"id": "e02bfa6c-8f0d-4413-8dcf-e644c07792d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1410864732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9235eeeb6fa4ccf959ffbea456f3694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa8c21c5-93", "ovs_interfaceid": "fa8c21c5-9351-49d0-8080-e693c96d1348", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 940.915171] env[61970]: DEBUG oslo_concurrency.lockutils [req-0a9d5283-0429-4a00-9c55-d6c5f7137926 req-8391530e-aa55-4cd5-9107-660d498ddc87 service nova] Acquired lock "refresh_cache-0d21db81-a53f-4edf-a3a6-f1949cba2499" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.915171] env[61970]: DEBUG nova.network.neutron [req-0a9d5283-0429-4a00-9c55-d6c5f7137926 req-8391530e-aa55-4cd5-9107-660d498ddc87 service nova] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Refreshing network info cache for port fa8c21c5-9351-49d0-8080-e693c96d1348 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 940.915610] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:cf:ec', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '510d3c47-3615-43d5-aa5d-a279fd915e71', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fa8c21c5-9351-49d0-8080-e693c96d1348', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 940.924730] env[61970]: DEBUG oslo.service.loopingcall [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.927763] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 940.928320] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ff84a351-b857-4e5b-88a6-930afe58709c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.949251] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 940.949251] env[61970]: value = "task-1355881" [ 940.949251] env[61970]: _type = "Task" [ 940.949251] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.957646] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355881, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.998276] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355877, 'name': Rename_Task, 'duration_secs': 0.141233} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.998624] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 940.998910] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3cb27af9-4a64-4a22-8395-8f95517147ba {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.005128] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 941.005128] env[61970]: value = "task-1355882" [ 941.005128] env[61970]: _type = "Task" [ 941.005128] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.013563] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355882, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.037729] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523b7dad-05b7-cdec-54e2-50b37544a256, 'name': SearchDatastore_Task, 'duration_secs': 0.00973} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.037986] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.038292] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 941.038563] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.038725] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.038909] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 941.039194] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6d94892a-04e8-40f5-bcc5-412b772059ea {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.048056] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 941.048246] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 941.049033] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a150d6e5-2808-4554-8fa1-fc9159e9cfed {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.054126] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 941.054126] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52247de9-d0f2-6076-c4d2-5a1313b5ba33" [ 941.054126] env[61970]: _type = "Task" [ 941.054126] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.062182] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52247de9-d0f2-6076-c4d2-5a1313b5ba33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.077366] env[61970]: INFO nova.compute.manager [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Took 17.53 seconds to build instance. [ 941.260959] env[61970]: DEBUG nova.network.neutron [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Successfully updated port: 490f6a87-12ad-4d7f-9dc5-46d694adf506 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 941.333615] env[61970]: DEBUG nova.network.neutron [req-0a9d5283-0429-4a00-9c55-d6c5f7137926 req-8391530e-aa55-4cd5-9107-660d498ddc87 service nova] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Updated VIF entry in instance network info cache for port fa8c21c5-9351-49d0-8080-e693c96d1348. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 941.334157] env[61970]: DEBUG nova.network.neutron [req-0a9d5283-0429-4a00-9c55-d6c5f7137926 req-8391530e-aa55-4cd5-9107-660d498ddc87 service nova] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Updating instance_info_cache with network_info: [{"id": "fa8c21c5-9351-49d0-8080-e693c96d1348", "address": "fa:16:3e:ce:cf:ec", "network": {"id": "e02bfa6c-8f0d-4413-8dcf-e644c07792d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1410864732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9235eeeb6fa4ccf959ffbea456f3694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa8c21c5-93", "ovs_interfaceid": "fa8c21c5-9351-49d0-8080-e693c96d1348", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.422140] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.422426] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.460326] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355881, 'name': CreateVM_Task, 'duration_secs': 0.320856} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.460500] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 941.461249] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.461475] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.461806] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 941.462082] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de04685a-d3d7-4107-b0cf-940fef5af923 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.467310] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 941.467310] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5293bac7-5ade-c61c-df58-ac83b3d9f02f" [ 941.467310] env[61970]: _type = "Task" [ 941.467310] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.475783] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5293bac7-5ade-c61c-df58-ac83b3d9f02f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.514761] env[61970]: DEBUG oslo_vmware.api [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355882, 'name': PowerOnVM_Task, 'duration_secs': 0.468214} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.515106] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 941.515259] env[61970]: INFO nova.compute.manager [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Took 7.79 seconds to spawn the instance on the hypervisor. [ 941.515444] env[61970]: DEBUG nova.compute.manager [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 941.516303] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7712aff-5eb3-4862-ba6e-372e6d39c5b2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.564542] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52247de9-d0f2-6076-c4d2-5a1313b5ba33, 'name': SearchDatastore_Task, 'duration_secs': 0.008423} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.566048] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b32c1503-1259-4a15-828f-054ef49d41f9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.572530] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 941.572530] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52184ff2-0388-54a3-db4d-3ba84b95c312" [ 941.572530] env[61970]: _type = "Task" [ 941.572530] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.581099] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2263812d-7ba3-4ec3-a386-eeb31c5e20da tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.042s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.585315] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52184ff2-0388-54a3-db4d-3ba84b95c312, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.766793] env[61970]: DEBUG oslo_concurrency.lockutils [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "refresh_cache-b9be8292-378b-4187-85dc-8d01e817faa3" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.766938] env[61970]: DEBUG oslo_concurrency.lockutils [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "refresh_cache-b9be8292-378b-4187-85dc-8d01e817faa3" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.767153] env[61970]: DEBUG nova.network.neutron [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 941.837750] env[61970]: DEBUG oslo_concurrency.lockutils [req-0a9d5283-0429-4a00-9c55-d6c5f7137926 req-8391530e-aa55-4cd5-9107-660d498ddc87 service nova] Releasing lock "refresh_cache-0d21db81-a53f-4edf-a3a6-f1949cba2499" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.865429] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.865787] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.865974] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Starting heal instance info cache {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 941.927596] env[61970]: INFO nova.compute.claims [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 941.985355] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5293bac7-5ade-c61c-df58-ac83b3d9f02f, 'name': SearchDatastore_Task, 'duration_secs': 0.010903} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.985671] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.985907] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 941.986137] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.038552] env[61970]: INFO nova.compute.manager [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Took 17.73 seconds to build instance. [ 942.084520] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52184ff2-0388-54a3-db4d-3ba84b95c312, 'name': SearchDatastore_Task, 'duration_secs': 0.020242} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.084816] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.085255] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] c4ec1797-d939-4c24-9314-ce9c74eb5aa1/c4ec1797-d939-4c24-9314-ce9c74eb5aa1.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 942.085538] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.085814] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 942.086318] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-152f641e-8d36-49ca-a819-ff19b810dd9c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.088706] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9b7019ed-2b8b-49f8-8dcc-e0fb4d465288 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.094913] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 942.094913] env[61970]: value = "task-1355883" [ 942.094913] env[61970]: _type = "Task" [ 942.094913] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.099020] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 942.099277] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 942.099918] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3aa153f0-ffac-4877-86c6-a75b738df433 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.106415] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355883, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.109770] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 942.109770] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528d6d43-2f98-44b8-19c3-7d2bc2f337fa" [ 942.109770] env[61970]: _type = "Task" [ 942.109770] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.115036] env[61970]: DEBUG nova.compute.manager [req-52079987-fb09-4c91-8b41-d007aa6c92ca req-06be03df-365c-414b-92b3-90a0c0877a9d service nova] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Received event network-changed-28a89dd6-da27-49dd-a87c-a702154014c1 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.115036] env[61970]: DEBUG nova.compute.manager [req-52079987-fb09-4c91-8b41-d007aa6c92ca req-06be03df-365c-414b-92b3-90a0c0877a9d service nova] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Refreshing instance network info cache due to event network-changed-28a89dd6-da27-49dd-a87c-a702154014c1. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 942.115207] env[61970]: DEBUG oslo_concurrency.lockutils [req-52079987-fb09-4c91-8b41-d007aa6c92ca req-06be03df-365c-414b-92b3-90a0c0877a9d service nova] Acquiring lock "refresh_cache-ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.115359] env[61970]: DEBUG oslo_concurrency.lockutils [req-52079987-fb09-4c91-8b41-d007aa6c92ca req-06be03df-365c-414b-92b3-90a0c0877a9d service nova] Acquired lock "refresh_cache-ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.115551] env[61970]: DEBUG nova.network.neutron [req-52079987-fb09-4c91-8b41-d007aa6c92ca req-06be03df-365c-414b-92b3-90a0c0877a9d service nova] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Refreshing network info cache for port 28a89dd6-da27-49dd-a87c-a702154014c1 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 942.123235] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528d6d43-2f98-44b8-19c3-7d2bc2f337fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.346618] env[61970]: DEBUG nova.network.neutron [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 942.441024] env[61970]: INFO nova.compute.resource_tracker [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Updating resource usage from migration a24105dc-3e41-4a23-aea8-b524dd9b68a8 [ 942.481993] env[61970]: DEBUG oslo_concurrency.lockutils [None req-11c42eb5-4aa1-4f61-b415-292bd7de1ada tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "fa8bf4b0-6955-4932-a85d-5a33947ae594" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.540515] env[61970]: DEBUG oslo_concurrency.lockutils [None req-302d5de2-d17e-46c4-99d2-82eec57361b8 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "fa8bf4b0-6955-4932-a85d-5a33947ae594" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.246s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.540946] env[61970]: DEBUG oslo_concurrency.lockutils [None req-11c42eb5-4aa1-4f61-b415-292bd7de1ada tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "fa8bf4b0-6955-4932-a85d-5a33947ae594" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.060s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.541202] env[61970]: DEBUG nova.compute.manager [None req-11c42eb5-4aa1-4f61-b415-292bd7de1ada tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 942.542401] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b17de6e-f00f-4c39-80ab-91ab676a67db {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.554200] env[61970]: DEBUG nova.compute.manager [None req-11c42eb5-4aa1-4f61-b415-292bd7de1ada tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61970) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 942.555270] env[61970]: DEBUG nova.objects.instance [None req-11c42eb5-4aa1-4f61-b415-292bd7de1ada tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lazy-loading 'flavor' on Instance uuid fa8bf4b0-6955-4932-a85d-5a33947ae594 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.606783] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355883, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.611996] env[61970]: DEBUG nova.network.neutron [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Updating instance_info_cache with network_info: [{"id": "490f6a87-12ad-4d7f-9dc5-46d694adf506", "address": "fa:16:3e:28:5d:e8", "network": {"id": "62edd125-b77c-4418-a465-1dcb3a0733c1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-371307366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6a4997482a4834a30c79c4cbab2cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap490f6a87-12", "ovs_interfaceid": "490f6a87-12ad-4d7f-9dc5-46d694adf506", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.626580] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528d6d43-2f98-44b8-19c3-7d2bc2f337fa, 'name': SearchDatastore_Task, 'duration_secs': 0.014664} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.628085] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00d152bf-398e-45e1-9f40-db9c80371448 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.641448] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 942.641448] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525e9d30-1a02-45f1-afe7-2c321a3df018" [ 942.641448] env[61970]: _type = "Task" [ 942.641448] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.646764] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525e9d30-1a02-45f1-afe7-2c321a3df018, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.702197] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015d2c4a-9cec-4b44-a957-a888e7b1b581 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.709442] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ca88ab-a093-448b-9525-d1b0355c2368 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.744084] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44fad831-a082-45cc-a6d0-c2630e0a4250 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.751788] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f5b9ca4-90b8-4554-9484-29925c985758 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.766150] env[61970]: DEBUG nova.compute.provider_tree [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.898381] env[61970]: DEBUG nova.compute.manager [req-e29f8fb6-d5f8-414f-adee-972a669b624c req-f5375954-4517-4ddf-8dad-c6662592cdfc service nova] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Received event network-vif-plugged-490f6a87-12ad-4d7f-9dc5-46d694adf506 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.898381] env[61970]: DEBUG oslo_concurrency.lockutils [req-e29f8fb6-d5f8-414f-adee-972a669b624c req-f5375954-4517-4ddf-8dad-c6662592cdfc service nova] Acquiring lock "b9be8292-378b-4187-85dc-8d01e817faa3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.898381] env[61970]: DEBUG oslo_concurrency.lockutils [req-e29f8fb6-d5f8-414f-adee-972a669b624c req-f5375954-4517-4ddf-8dad-c6662592cdfc service nova] Lock "b9be8292-378b-4187-85dc-8d01e817faa3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.898381] env[61970]: DEBUG oslo_concurrency.lockutils [req-e29f8fb6-d5f8-414f-adee-972a669b624c req-f5375954-4517-4ddf-8dad-c6662592cdfc service nova] Lock "b9be8292-378b-4187-85dc-8d01e817faa3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.898381] env[61970]: DEBUG nova.compute.manager [req-e29f8fb6-d5f8-414f-adee-972a669b624c req-f5375954-4517-4ddf-8dad-c6662592cdfc service nova] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] No waiting events found dispatching network-vif-plugged-490f6a87-12ad-4d7f-9dc5-46d694adf506 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 942.898381] env[61970]: WARNING nova.compute.manager [req-e29f8fb6-d5f8-414f-adee-972a669b624c req-f5375954-4517-4ddf-8dad-c6662592cdfc service nova] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Received unexpected event network-vif-plugged-490f6a87-12ad-4d7f-9dc5-46d694adf506 for instance with vm_state building and task_state spawning. [ 942.898381] env[61970]: DEBUG nova.compute.manager [req-e29f8fb6-d5f8-414f-adee-972a669b624c req-f5375954-4517-4ddf-8dad-c6662592cdfc service nova] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Received event network-changed-490f6a87-12ad-4d7f-9dc5-46d694adf506 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.898381] env[61970]: DEBUG nova.compute.manager [req-e29f8fb6-d5f8-414f-adee-972a669b624c req-f5375954-4517-4ddf-8dad-c6662592cdfc service nova] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Refreshing instance network info cache due to event network-changed-490f6a87-12ad-4d7f-9dc5-46d694adf506. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 942.898381] env[61970]: DEBUG oslo_concurrency.lockutils [req-e29f8fb6-d5f8-414f-adee-972a669b624c req-f5375954-4517-4ddf-8dad-c6662592cdfc service nova] Acquiring lock "refresh_cache-b9be8292-378b-4187-85dc-8d01e817faa3" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.941521] env[61970]: DEBUG nova.network.neutron [req-52079987-fb09-4c91-8b41-d007aa6c92ca req-06be03df-365c-414b-92b3-90a0c0877a9d service nova] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Updated VIF entry in instance network info cache for port 28a89dd6-da27-49dd-a87c-a702154014c1. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 942.941934] env[61970]: DEBUG nova.network.neutron [req-52079987-fb09-4c91-8b41-d007aa6c92ca req-06be03df-365c-414b-92b3-90a0c0877a9d service nova] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Updating instance_info_cache with network_info: [{"id": "28a89dd6-da27-49dd-a87c-a702154014c1", "address": "fa:16:3e:82:0a:58", "network": {"id": "42def9ae-4eed-41e5-a538-81eaced2cdd4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2047751481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19dae2316adb4dc980bd972c8c40c034", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c1b8689-a9b4-4972-beb9-6a1c8de1dc88", "external-id": "nsx-vlan-transportzone-455", "segmentation_id": 455, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28a89dd6-da", "ovs_interfaceid": "28a89dd6-da27-49dd-a87c-a702154014c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.062764] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-11c42eb5-4aa1-4f61-b415-292bd7de1ada tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 943.063051] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bb686744-767b-4bc9-a5cf-98bfa4455782 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.069829] env[61970]: DEBUG oslo_vmware.api [None req-11c42eb5-4aa1-4f61-b415-292bd7de1ada tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 943.069829] env[61970]: value = "task-1355885" [ 943.069829] env[61970]: _type = "Task" [ 943.069829] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.078384] env[61970]: DEBUG oslo_vmware.api [None req-11c42eb5-4aa1-4f61-b415-292bd7de1ada tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355885, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.105306] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355883, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.627531} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.107142] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] c4ec1797-d939-4c24-9314-ce9c74eb5aa1/c4ec1797-d939-4c24-9314-ce9c74eb5aa1.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 943.107142] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 943.107142] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-833a4d1f-0b53-411e-92b0-fe3c3e56136d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.112389] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 943.112389] env[61970]: value = "task-1355886" [ 943.112389] env[61970]: _type = "Task" [ 943.112389] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.121814] env[61970]: DEBUG oslo_concurrency.lockutils [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "refresh_cache-b9be8292-378b-4187-85dc-8d01e817faa3" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.122212] env[61970]: DEBUG nova.compute.manager [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Instance network_info: |[{"id": "490f6a87-12ad-4d7f-9dc5-46d694adf506", "address": "fa:16:3e:28:5d:e8", "network": {"id": "62edd125-b77c-4418-a465-1dcb3a0733c1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-371307366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6a4997482a4834a30c79c4cbab2cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap490f6a87-12", "ovs_interfaceid": "490f6a87-12ad-4d7f-9dc5-46d694adf506", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 943.122625] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355886, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.122979] env[61970]: DEBUG oslo_concurrency.lockutils [req-e29f8fb6-d5f8-414f-adee-972a669b624c req-f5375954-4517-4ddf-8dad-c6662592cdfc service nova] Acquired lock "refresh_cache-b9be8292-378b-4187-85dc-8d01e817faa3" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.123304] env[61970]: DEBUG nova.network.neutron [req-e29f8fb6-d5f8-414f-adee-972a669b624c req-f5375954-4517-4ddf-8dad-c6662592cdfc service nova] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Refreshing network info cache for port 490f6a87-12ad-4d7f-9dc5-46d694adf506 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 943.124831] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:28:5d:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '490f6a87-12ad-4d7f-9dc5-46d694adf506', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 943.133165] env[61970]: DEBUG oslo.service.loopingcall [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 943.136438] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 943.137074] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fb9cb1cd-3f50-49f6-a53a-fec933e2b970 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.167178] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525e9d30-1a02-45f1-afe7-2c321a3df018, 'name': SearchDatastore_Task, 'duration_secs': 0.065741} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.167178] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.167178] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] 0d21db81-a53f-4edf-a3a6-f1949cba2499/0d21db81-a53f-4edf-a3a6-f1949cba2499.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 943.167816] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 943.167816] env[61970]: value = "task-1355887" [ 943.167816] env[61970]: _type = "Task" [ 943.167816] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.167816] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c5dd043e-8d09-461f-8114-8d5f1cc54d10 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.178419] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355887, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.180168] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 943.180168] env[61970]: value = "task-1355888" [ 943.180168] env[61970]: _type = "Task" [ 943.180168] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.193956] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355888, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.269719] env[61970]: DEBUG nova.scheduler.client.report [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 943.386649] env[61970]: DEBUG nova.network.neutron [req-e29f8fb6-d5f8-414f-adee-972a669b624c req-f5375954-4517-4ddf-8dad-c6662592cdfc service nova] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Updated VIF entry in instance network info cache for port 490f6a87-12ad-4d7f-9dc5-46d694adf506. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 943.387030] env[61970]: DEBUG nova.network.neutron [req-e29f8fb6-d5f8-414f-adee-972a669b624c req-f5375954-4517-4ddf-8dad-c6662592cdfc service nova] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Updating instance_info_cache with network_info: [{"id": "490f6a87-12ad-4d7f-9dc5-46d694adf506", "address": "fa:16:3e:28:5d:e8", "network": {"id": "62edd125-b77c-4418-a465-1dcb3a0733c1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-371307366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6a4997482a4834a30c79c4cbab2cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap490f6a87-12", "ovs_interfaceid": "490f6a87-12ad-4d7f-9dc5-46d694adf506", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.444421] env[61970]: DEBUG oslo_concurrency.lockutils [req-52079987-fb09-4c91-8b41-d007aa6c92ca req-06be03df-365c-414b-92b3-90a0c0877a9d service nova] Releasing lock "refresh_cache-ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.579537] env[61970]: DEBUG oslo_vmware.api [None req-11c42eb5-4aa1-4f61-b415-292bd7de1ada tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355885, 'name': PowerOffVM_Task, 'duration_secs': 0.206625} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.579808] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-11c42eb5-4aa1-4f61-b415-292bd7de1ada tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 943.579986] env[61970]: DEBUG nova.compute.manager [None req-11c42eb5-4aa1-4f61-b415-292bd7de1ada tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 943.580795] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee5969ef-2c90-41b8-8724-04bb8389d3b4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.622516] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355886, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071864} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.622757] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 943.623606] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4181d86-66fb-4d89-98f1-b314a04778c7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.647145] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] c4ec1797-d939-4c24-9314-ce9c74eb5aa1/c4ec1797-d939-4c24-9314-ce9c74eb5aa1.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.647822] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31e3f110-0868-464b-8964-b12596ab9335 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.668278] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 943.668278] env[61970]: value = "task-1355889" [ 943.668278] env[61970]: _type = "Task" [ 943.668278] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.683365] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355887, 'name': CreateVM_Task} progress is 25%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.683750] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355889, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.693166] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355888, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.777219] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.354s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.777457] env[61970]: INFO nova.compute.manager [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Migrating [ 943.889971] env[61970]: DEBUG oslo_concurrency.lockutils [req-e29f8fb6-d5f8-414f-adee-972a669b624c req-f5375954-4517-4ddf-8dad-c6662592cdfc service nova] Releasing lock "refresh_cache-b9be8292-378b-4187-85dc-8d01e817faa3" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.093996] env[61970]: DEBUG oslo_concurrency.lockutils [None req-11c42eb5-4aa1-4f61-b415-292bd7de1ada tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "fa8bf4b0-6955-4932-a85d-5a33947ae594" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.552s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.184026] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355889, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.186147] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355887, 'name': CreateVM_Task, 'duration_secs': 0.913102} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.192024] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 944.192024] env[61970]: DEBUG oslo_concurrency.lockutils [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.192024] env[61970]: DEBUG oslo_concurrency.lockutils [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.192024] env[61970]: DEBUG oslo_concurrency.lockutils [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 944.192024] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4104a8cd-8d51-482f-895f-181f811071a0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.197281] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355888, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.710757} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.197977] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] 0d21db81-a53f-4edf-a3a6-f1949cba2499/0d21db81-a53f-4edf-a3a6-f1949cba2499.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 944.198426] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 944.199254] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ae42e70e-1acd-416f-ab43-e7fe658fd6b5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.204017] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 944.204017] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52183e29-095b-81c8-8043-f3b2ef58cd30" [ 944.204017] env[61970]: _type = "Task" [ 944.204017] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.208330] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 944.208330] env[61970]: value = "task-1355890" [ 944.208330] env[61970]: _type = "Task" [ 944.208330] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.212238] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52183e29-095b-81c8-8043-f3b2ef58cd30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.219661] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355890, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.298907] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "refresh_cache-dc38f9da-9f82-4707-9764-b28aa83a8ece" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.298907] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "refresh_cache-dc38f9da-9f82-4707-9764-b28aa83a8ece" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.298907] env[61970]: DEBUG nova.network.neutron [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 944.383567] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Didn't find any instances for network info cache update. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 944.383810] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 944.384092] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 944.384298] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 944.384493] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 944.384687] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 944.384985] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 944.385136] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61970) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 944.385286] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 944.679458] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355889, 'name': ReconfigVM_Task, 'duration_secs': 0.790239} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.679458] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Reconfigured VM instance instance-00000057 to attach disk [datastore1] c4ec1797-d939-4c24-9314-ce9c74eb5aa1/c4ec1797-d939-4c24-9314-ce9c74eb5aa1.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.680354] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5722c624-768e-4e6f-8043-c468fe17cd95 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.686711] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 944.686711] env[61970]: value = "task-1355891" [ 944.686711] env[61970]: _type = "Task" [ 944.686711] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.695231] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355891, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.717010] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52183e29-095b-81c8-8043-f3b2ef58cd30, 'name': SearchDatastore_Task, 'duration_secs': 0.009318} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.721691] env[61970]: DEBUG oslo_concurrency.lockutils [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.722017] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 944.722350] env[61970]: DEBUG oslo_concurrency.lockutils [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.722579] env[61970]: DEBUG oslo_concurrency.lockutils [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.722843] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 944.723204] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-584fdb12-6921-4475-b5c9-9fbed4ded151 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.731075] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355890, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06714} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.731408] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 944.732518] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-488ae29d-68e4-44c5-8513-6e462c737457 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.737562] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 944.737562] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 944.738878] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-990438e2-fbe3-404a-8d43-df008c515dd5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.765054] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 0d21db81-a53f-4edf-a3a6-f1949cba2499/0d21db81-a53f-4edf-a3a6-f1949cba2499.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 944.765732] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9f70c06-33aa-4c05-83ec-0b5f0009a9cf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.781282] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 944.781282] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5250c3cd-4917-6220-535d-f85cc8a8f428" [ 944.781282] env[61970]: _type = "Task" [ 944.781282] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.786325] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 944.786325] env[61970]: value = "task-1355892" [ 944.786325] env[61970]: _type = "Task" [ 944.786325] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.793753] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5250c3cd-4917-6220-535d-f85cc8a8f428, 'name': SearchDatastore_Task, 'duration_secs': 0.009473} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.794743] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3563733-8b68-41cf-9383-e2f98e89e766 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.799342] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355892, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.803564] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 944.803564] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]521b0a0d-7949-2fdc-859b-6100b61a33dd" [ 944.803564] env[61970]: _type = "Task" [ 944.803564] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.811026] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]521b0a0d-7949-2fdc-859b-6100b61a33dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.887828] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.888081] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.888249] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.888406] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61970) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 944.889382] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8606632-2662-468f-a224-178d96dbb4e4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.897610] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472154bb-f890-4813-8ff8-22146bddc452 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.912171] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Volume attach. Driver type: vmdk {{(pid=61970) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 944.912404] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288840', 'volume_id': '3916ae0c-f673-4a6a-bd2a-9134a92569c0', 'name': 'volume-3916ae0c-f673-4a6a-bd2a-9134a92569c0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd0ad9689-bdb7-4d68-b19a-3f805ab4612e', 'attached_at': '', 'detached_at': '', 'volume_id': '3916ae0c-f673-4a6a-bd2a-9134a92569c0', 'serial': '3916ae0c-f673-4a6a-bd2a-9134a92569c0'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 944.913206] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e024aa8-e301-4d22-be87-0a3c28ae66f9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.916103] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef703130-b5d6-41d2-90a6-45f490c676a1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.935246] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b577fe-5ae9-4633-a4aa-73e13bf8333a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.941889] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea9ea37c-5729-420f-800a-135ecc27bf6f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.967178] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] volume-3916ae0c-f673-4a6a-bd2a-9134a92569c0/volume-3916ae0c-f673-4a6a-bd2a-9134a92569c0.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 944.999264] env[61970]: INFO nova.compute.manager [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Rebuilding instance [ 944.999264] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d3c005c-59dd-457d-b3f7-6a0d78ecb90a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.012414] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180827MB free_disk=150GB free_vcpus=48 pci_devices=None {{(pid=61970) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 945.012576] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.012770] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.022792] env[61970]: DEBUG oslo_vmware.api [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 945.022792] env[61970]: value = "task-1355893" [ 945.022792] env[61970]: _type = "Task" [ 945.022792] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.031115] env[61970]: DEBUG oslo_vmware.api [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355893, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.058629] env[61970]: DEBUG nova.compute.manager [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.059512] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f95f908-ab10-4832-8a03-5558de44736a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.100420] env[61970]: DEBUG nova.network.neutron [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Updating instance_info_cache with network_info: [{"id": "313663a2-7d1f-40ea-abfc-c540d35235ca", "address": "fa:16:3e:c7:5e:f0", "network": {"id": "cba185ff-bfa3-4547-b7b0-638c3e79a24c", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-864852491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5c64cb9232fb413cbd7627dcf077e9ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap313663a2-7d", "ovs_interfaceid": "313663a2-7d1f-40ea-abfc-c540d35235ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.196626] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355891, 'name': Rename_Task, 'duration_secs': 0.28405} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.196626] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 945.196816] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-accf491c-20e4-4054-ac78-ad940f15a56a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.203462] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 945.203462] env[61970]: value = "task-1355894" [ 945.203462] env[61970]: _type = "Task" [ 945.203462] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.210952] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355894, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.298029] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355892, 'name': ReconfigVM_Task, 'duration_secs': 0.27557} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.298029] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 0d21db81-a53f-4edf-a3a6-f1949cba2499/0d21db81-a53f-4edf-a3a6-f1949cba2499.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 945.298530] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b8cda516-8857-450c-86fc-b42e4e368c98 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.309267] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 945.309267] env[61970]: value = "task-1355895" [ 945.309267] env[61970]: _type = "Task" [ 945.309267] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.315521] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]521b0a0d-7949-2fdc-859b-6100b61a33dd, 'name': SearchDatastore_Task, 'duration_secs': 0.014032} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.316075] env[61970]: DEBUG oslo_concurrency.lockutils [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.316355] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] b9be8292-378b-4187-85dc-8d01e817faa3/b9be8292-378b-4187-85dc-8d01e817faa3.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 945.316610] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a9dca9de-f521-43cb-9a2a-b3eaaeb96f31 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.321390] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355895, 'name': Rename_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.325629] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 945.325629] env[61970]: value = "task-1355896" [ 945.325629] env[61970]: _type = "Task" [ 945.325629] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.333063] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355896, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.533518] env[61970]: DEBUG oslo_vmware.api [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355893, 'name': ReconfigVM_Task, 'duration_secs': 0.35979} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.533808] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Reconfigured VM instance instance-0000004e to attach disk [datastore1] volume-3916ae0c-f673-4a6a-bd2a-9134a92569c0/volume-3916ae0c-f673-4a6a-bd2a-9134a92569c0.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 945.539443] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2658f8c0-a34b-46f4-993b-60cfe3b68f09 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.573177] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 945.573525] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d345859-63f5-4436-9e8a-81cf1ff383d5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.603641] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "refresh_cache-dc38f9da-9f82-4707-9764-b28aa83a8ece" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.619340] env[61970]: DEBUG oslo_vmware.api [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 945.619340] env[61970]: value = "task-1355897" [ 945.619340] env[61970]: _type = "Task" [ 945.619340] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.621094] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 945.621094] env[61970]: value = "task-1355898" [ 945.621094] env[61970]: _type = "Task" [ 945.621094] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.636736] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] VM already powered off {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 945.636963] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 945.637276] env[61970]: DEBUG oslo_vmware.api [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355897, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.638023] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e082ad41-40a3-4947-af46-56f7b1dd219d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.645641] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 945.645931] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4b5ffe1c-7ef9-4105-9432-4253fadd98e9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.715506] env[61970]: DEBUG oslo_vmware.api [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355894, 'name': PowerOnVM_Task, 'duration_secs': 0.452951} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.715743] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 945.715953] env[61970]: INFO nova.compute.manager [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Took 9.73 seconds to spawn the instance on the hypervisor. [ 945.716190] env[61970]: DEBUG nova.compute.manager [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.717055] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c555c5e1-cb1d-4e80-acad-1df92c8940f3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.720897] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 945.721138] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 945.721328] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleting the datastore file [datastore2] fa8bf4b0-6955-4932-a85d-5a33947ae594 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 945.721962] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f02e996a-9c4b-4727-acbc-3ccaabd32fd0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.732502] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 945.732502] env[61970]: value = "task-1355900" [ 945.732502] env[61970]: _type = "Task" [ 945.732502] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.740680] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355900, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.818814] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355895, 'name': Rename_Task, 'duration_secs': 0.12955} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.819216] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 945.819595] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-05d3da01-160d-4daf-b577-94b22fbf1865 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.825390] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 945.825390] env[61970]: value = "task-1355901" [ 945.825390] env[61970]: _type = "Task" [ 945.825390] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.836199] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355896, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496913} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.839055] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] b9be8292-378b-4187-85dc-8d01e817faa3/b9be8292-378b-4187-85dc-8d01e817faa3.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 945.839293] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 945.839565] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355901, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.839774] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ec62b6dd-8e1e-4d12-96ac-96ddaec2de79 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.845895] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 945.845895] env[61970]: value = "task-1355902" [ 945.845895] env[61970]: _type = "Task" [ 945.845895] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.853474] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355902, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.025289] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Applying migration context for instance dc38f9da-9f82-4707-9764-b28aa83a8ece as it has an incoming, in-progress migration a24105dc-3e41-4a23-aea8-b524dd9b68a8. Migration status is pre-migrating {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 946.026407] env[61970]: INFO nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Updating resource usage from migration a24105dc-3e41-4a23-aea8-b524dd9b68a8 [ 946.045272] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 946.045440] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 0304640a-6c46-4f57-a8b8-5058da2e3de4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 946.045772] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance d0ad9689-bdb7-4d68-b19a-3f805ab4612e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 946.045772] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 7f59000a-94d9-45b6-aa7a-300d95793615 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 946.045931] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance ebb92a5b-f635-4504-a57f-395de514015a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 946.045992] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 3b346c68-53a9-4ceb-83a3-9e4bce9610b1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 946.046119] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 946.046252] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance fa8bf4b0-6955-4932-a85d-5a33947ae594 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 946.046506] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance c4ec1797-d939-4c24-9314-ce9c74eb5aa1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 946.046879] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 0d21db81-a53f-4edf-a3a6-f1949cba2499 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 946.046879] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance b9be8292-378b-4187-85dc-8d01e817faa3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 946.047047] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Migration a24105dc-3e41-4a23-aea8-b524dd9b68a8 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 946.047111] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance dc38f9da-9f82-4707-9764-b28aa83a8ece actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 946.047306] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 946.047444] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3072MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 946.131867] env[61970]: DEBUG oslo_vmware.api [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355897, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.207580] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c21314ec-7547-4cca-b2fb-2242d564e873 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.216093] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c9c559-4dfd-4795-a908-110c72e5900b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.253569] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d427fd2e-8197-4b9b-a457-52133ce08b6b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.257989] env[61970]: INFO nova.compute.manager [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Took 19.46 seconds to build instance. [ 946.265921] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355900, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.222472} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.266283] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.266476] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 946.266655] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 946.270309] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1251405-a0bb-4ac8-9567-a49c81351445 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.284708] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.337394] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355901, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.356885] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355902, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06246} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.357188] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 946.358101] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4ea8f9-a6ca-48ad-b35f-b9ac40e2f585 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.380967] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] b9be8292-378b-4187-85dc-8d01e817faa3/b9be8292-378b-4187-85dc-8d01e817faa3.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 946.381287] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b7c566e-e120-49a4-af5a-0a4b54552810 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.401070] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 946.401070] env[61970]: value = "task-1355903" [ 946.401070] env[61970]: _type = "Task" [ 946.401070] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.408813] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355903, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.634304] env[61970]: DEBUG oslo_vmware.api [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355897, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.764995] env[61970]: DEBUG oslo_concurrency.lockutils [None req-62704198-4ca1-4c1d-819d-b513c46ef171 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.978s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.787582] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 946.836756] env[61970]: DEBUG oslo_vmware.api [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355901, 'name': PowerOnVM_Task, 'duration_secs': 0.549003} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.837111] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 946.837344] env[61970]: INFO nova.compute.manager [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Took 8.57 seconds to spawn the instance on the hypervisor. [ 946.838081] env[61970]: DEBUG nova.compute.manager [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 946.839020] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44308ff6-ca8d-4a83-93f7-fffb37608099 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.865529] env[61970]: DEBUG nova.compute.manager [req-d463ad97-cb4c-4c26-a5dd-cc51d11d68e9 req-9008176c-6ca8-40b7-81ac-b49bbe06bc65 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Received event network-changed-5b506f12-7feb-41e2-8e18-abc9262ac87b {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.865764] env[61970]: DEBUG nova.compute.manager [req-d463ad97-cb4c-4c26-a5dd-cc51d11d68e9 req-9008176c-6ca8-40b7-81ac-b49bbe06bc65 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Refreshing instance network info cache due to event network-changed-5b506f12-7feb-41e2-8e18-abc9262ac87b. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 946.865988] env[61970]: DEBUG oslo_concurrency.lockutils [req-d463ad97-cb4c-4c26-a5dd-cc51d11d68e9 req-9008176c-6ca8-40b7-81ac-b49bbe06bc65 service nova] Acquiring lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.866144] env[61970]: DEBUG oslo_concurrency.lockutils [req-d463ad97-cb4c-4c26-a5dd-cc51d11d68e9 req-9008176c-6ca8-40b7-81ac-b49bbe06bc65 service nova] Acquired lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.866304] env[61970]: DEBUG nova.network.neutron [req-d463ad97-cb4c-4c26-a5dd-cc51d11d68e9 req-9008176c-6ca8-40b7-81ac-b49bbe06bc65 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Refreshing network info cache for port 5b506f12-7feb-41e2-8e18-abc9262ac87b {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 946.913449] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.120268] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0b98c4-3654-4b9b-809c-d0d489ebd292 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.140518] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Updating instance 'dc38f9da-9f82-4707-9764-b28aa83a8ece' progress to 0 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 947.152031] env[61970]: DEBUG oslo_vmware.api [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355897, 'name': ReconfigVM_Task, 'duration_secs': 1.249629} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.152031] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288840', 'volume_id': '3916ae0c-f673-4a6a-bd2a-9134a92569c0', 'name': 'volume-3916ae0c-f673-4a6a-bd2a-9134a92569c0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd0ad9689-bdb7-4d68-b19a-3f805ab4612e', 'attached_at': '', 'detached_at': '', 'volume_id': '3916ae0c-f673-4a6a-bd2a-9134a92569c0', 'serial': '3916ae0c-f673-4a6a-bd2a-9134a92569c0'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 947.295035] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 947.295247] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.282s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.306855] env[61970]: DEBUG nova.virt.hardware [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 947.307128] env[61970]: DEBUG nova.virt.hardware [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 947.307294] env[61970]: DEBUG nova.virt.hardware [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 947.307523] env[61970]: DEBUG nova.virt.hardware [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 947.307625] env[61970]: DEBUG nova.virt.hardware [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 947.307783] env[61970]: DEBUG nova.virt.hardware [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 947.308173] env[61970]: DEBUG nova.virt.hardware [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 947.308409] env[61970]: DEBUG nova.virt.hardware [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 947.308654] env[61970]: DEBUG nova.virt.hardware [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 947.308858] env[61970]: DEBUG nova.virt.hardware [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 947.309053] env[61970]: DEBUG nova.virt.hardware [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 947.310087] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0fbd6c-5bc9-4cf7-966c-a166f79afcc6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.319064] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff714263-dee1-4882-98d0-c89f1c1022ec {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.334154] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1c:34:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b356db78-99c7-4464-822c-fc7e193f7878', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'df21eb74-2d16-49a2-9a94-372df5b2f8a7', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 947.341820] env[61970]: DEBUG oslo.service.loopingcall [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 947.342165] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 947.342341] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2084991b-0fd2-42ed-92a9-57872ab360ee {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.364903] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 947.364903] env[61970]: value = "task-1355904" [ 947.364903] env[61970]: _type = "Task" [ 947.364903] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.365416] env[61970]: INFO nova.compute.manager [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Took 18.66 seconds to build instance. [ 947.377247] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355904, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.412330] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355903, 'name': ReconfigVM_Task, 'duration_secs': 0.909607} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.413364] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Reconfigured VM instance instance-00000059 to attach disk [datastore1] b9be8292-378b-4187-85dc-8d01e817faa3/b9be8292-378b-4187-85dc-8d01e817faa3.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 947.414160] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ff736d8a-4d54-47d8-835a-a7f63965301c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.419913] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 947.419913] env[61970]: value = "task-1355905" [ 947.419913] env[61970]: _type = "Task" [ 947.419913] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.431766] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355905, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.623266] env[61970]: DEBUG nova.network.neutron [req-d463ad97-cb4c-4c26-a5dd-cc51d11d68e9 req-9008176c-6ca8-40b7-81ac-b49bbe06bc65 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Updated VIF entry in instance network info cache for port 5b506f12-7feb-41e2-8e18-abc9262ac87b. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 947.623722] env[61970]: DEBUG nova.network.neutron [req-d463ad97-cb4c-4c26-a5dd-cc51d11d68e9 req-9008176c-6ca8-40b7-81ac-b49bbe06bc65 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Updating instance_info_cache with network_info: [{"id": "5b506f12-7feb-41e2-8e18-abc9262ac87b", "address": "fa:16:3e:f9:43:05", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b506f12-7f", "ovs_interfaceid": "5b506f12-7feb-41e2-8e18-abc9262ac87b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.647062] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 947.648474] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6d7bdee7-2782-4dd8-afc2-cfa0e7d700b1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.653151] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.653407] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.661564] env[61970]: DEBUG oslo_vmware.api [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 947.661564] env[61970]: value = "task-1355906" [ 947.661564] env[61970]: _type = "Task" [ 947.661564] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.673594] env[61970]: DEBUG oslo_vmware.api [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355906, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.870822] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7f54e5e-6423-4d61-a0a7-6e95c4de5c07 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "0d21db81-a53f-4edf-a3a6-f1949cba2499" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.174s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.876609] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355904, 'name': CreateVM_Task, 'duration_secs': 0.421441} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.876669] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 947.877365] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.877454] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.877808] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 947.878045] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7799f91a-04f5-4dab-acff-0984b6453a9f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.883267] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 947.883267] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5273604d-15b9-abe1-2334-491fff0dfc63" [ 947.883267] env[61970]: _type = "Task" [ 947.883267] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.892331] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5273604d-15b9-abe1-2334-491fff0dfc63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.929476] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355905, 'name': Rename_Task, 'duration_secs': 0.233809} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.929788] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 947.930072] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab8a6734-3ee8-44a2-8683-f6f9a9bf8ef0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.936674] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 947.936674] env[61970]: value = "task-1355907" [ 947.936674] env[61970]: _type = "Task" [ 947.936674] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.946496] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355907, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.006984] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "0d21db81-a53f-4edf-a3a6-f1949cba2499" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.007271] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "0d21db81-a53f-4edf-a3a6-f1949cba2499" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.007515] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "0d21db81-a53f-4edf-a3a6-f1949cba2499-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.007724] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "0d21db81-a53f-4edf-a3a6-f1949cba2499-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.007927] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "0d21db81-a53f-4edf-a3a6-f1949cba2499-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.010739] env[61970]: INFO nova.compute.manager [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Terminating instance [ 948.013681] env[61970]: DEBUG nova.compute.manager [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 948.013983] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 948.014932] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ffd1873-67b8-497b-8ce5-47cd32bb1724 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.023350] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 948.023632] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-75b7d16d-fd41-4d7f-bd48-266a23491a42 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.029769] env[61970]: DEBUG oslo_vmware.api [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 948.029769] env[61970]: value = "task-1355908" [ 948.029769] env[61970]: _type = "Task" [ 948.029769] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.037375] env[61970]: DEBUG oslo_vmware.api [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355908, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.127790] env[61970]: DEBUG oslo_concurrency.lockutils [req-d463ad97-cb4c-4c26-a5dd-cc51d11d68e9 req-9008176c-6ca8-40b7-81ac-b49bbe06bc65 service nova] Releasing lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.158461] env[61970]: DEBUG nova.compute.manager [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 948.171035] env[61970]: DEBUG oslo_vmware.api [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355906, 'name': PowerOffVM_Task, 'duration_secs': 0.358001} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.171343] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 948.171547] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Updating instance 'dc38f9da-9f82-4707-9764-b28aa83a8ece' progress to 17 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 948.199753] env[61970]: DEBUG nova.objects.instance [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lazy-loading 'flavor' on Instance uuid d0ad9689-bdb7-4d68-b19a-3f805ab4612e {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.395057] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5273604d-15b9-abe1-2334-491fff0dfc63, 'name': SearchDatastore_Task, 'duration_secs': 0.015484} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.395397] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.395653] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 948.396027] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.396215] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.396404] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 948.396704] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-20fba5f7-4cd6-4da8-af59-b58c5e4be12c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.405106] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 948.405367] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 948.406162] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-785ae2c6-14d8-4054-8d95-7f0cf04c421c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.411713] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 948.411713] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525656ba-4ae2-fa40-09df-b6891cd92689" [ 948.411713] env[61970]: _type = "Task" [ 948.411713] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.420216] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525656ba-4ae2-fa40-09df-b6891cd92689, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.446615] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355907, 'name': PowerOnVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.540172] env[61970]: DEBUG oslo_vmware.api [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355908, 'name': PowerOffVM_Task, 'duration_secs': 0.202338} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.540448] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 948.540621] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 948.540878] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af0300b9-7b96-4831-8a4a-38f18629f399 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.637570] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 948.637954] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 948.638238] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleting the datastore file [datastore1] 0d21db81-a53f-4edf-a3a6-f1949cba2499 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 948.638619] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-67723b66-a71a-4003-924b-a9c3cc54e08e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.646545] env[61970]: DEBUG oslo_vmware.api [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 948.646545] env[61970]: value = "task-1355910" [ 948.646545] env[61970]: _type = "Task" [ 948.646545] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.657079] env[61970]: DEBUG oslo_vmware.api [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355910, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.681250] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 948.681511] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 948.681675] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 948.681862] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 948.682024] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 948.682205] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 948.682384] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 948.682550] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 948.682719] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 948.682886] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 948.683071] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 948.689094] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.689339] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.690908] env[61970]: INFO nova.compute.claims [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 948.693480] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ced2000e-c979-4d8e-b4dc-36d640dcbebf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.706559] env[61970]: DEBUG oslo_concurrency.lockutils [None req-db24d6ad-31cd-4108-b872-b023e94ec093 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.421s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.711121] env[61970]: DEBUG oslo_vmware.api [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 948.711121] env[61970]: value = "task-1355911" [ 948.711121] env[61970]: _type = "Task" [ 948.711121] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.715148] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0f33a1f-83e8-493a-aaf9-e37fe62c6b84 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.715383] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0f33a1f-83e8-493a-aaf9-e37fe62c6b84 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.715569] env[61970]: DEBUG nova.compute.manager [None req-b0f33a1f-83e8-493a-aaf9-e37fe62c6b84 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 948.716434] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b64dbf-cca7-4a4e-a7c4-dc9bed0b55c6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.726844] env[61970]: DEBUG oslo_vmware.api [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355911, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.727187] env[61970]: DEBUG nova.compute.manager [None req-b0f33a1f-83e8-493a-aaf9-e37fe62c6b84 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61970) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 948.727748] env[61970]: DEBUG nova.objects.instance [None req-b0f33a1f-83e8-493a-aaf9-e37fe62c6b84 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lazy-loading 'flavor' on Instance uuid d0ad9689-bdb7-4d68-b19a-3f805ab4612e {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.922642] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525656ba-4ae2-fa40-09df-b6891cd92689, 'name': SearchDatastore_Task, 'duration_secs': 0.009678} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.923483] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ed9805d-c8b1-41bc-84e1-704e61d3005c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.928644] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 948.928644] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529b72d5-9718-1f2f-5b7a-8a4b14ee8909" [ 948.928644] env[61970]: _type = "Task" [ 948.928644] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.936971] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529b72d5-9718-1f2f-5b7a-8a4b14ee8909, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.945390] env[61970]: DEBUG oslo_vmware.api [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355907, 'name': PowerOnVM_Task, 'duration_secs': 0.551778} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.945689] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 948.945922] env[61970]: INFO nova.compute.manager [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Took 8.37 seconds to spawn the instance on the hypervisor. [ 948.946175] env[61970]: DEBUG nova.compute.manager [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 948.946891] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb9e3cc-4283-4325-b0e4-6a5a6d72dfed {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.157612] env[61970]: DEBUG oslo_vmware.api [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355910, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.221482] env[61970]: DEBUG oslo_vmware.api [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355911, 'name': ReconfigVM_Task, 'duration_secs': 0.313057} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.221832] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Updating instance 'dc38f9da-9f82-4707-9764-b28aa83a8ece' progress to 33 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 949.232410] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0f33a1f-83e8-493a-aaf9-e37fe62c6b84 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 949.232654] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2cfdc050-17bd-4f0a-9cda-2caff90fb566 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.239638] env[61970]: DEBUG oslo_vmware.api [None req-b0f33a1f-83e8-493a-aaf9-e37fe62c6b84 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 949.239638] env[61970]: value = "task-1355912" [ 949.239638] env[61970]: _type = "Task" [ 949.239638] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.247375] env[61970]: DEBUG oslo_vmware.api [None req-b0f33a1f-83e8-493a-aaf9-e37fe62c6b84 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355912, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.439271] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529b72d5-9718-1f2f-5b7a-8a4b14ee8909, 'name': SearchDatastore_Task, 'duration_secs': 0.011156} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.439557] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.439876] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] fa8bf4b0-6955-4932-a85d-5a33947ae594/fa8bf4b0-6955-4932-a85d-5a33947ae594.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 949.440268] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7245955f-9cde-4f36-a7bf-181e05a4a881 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.447097] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 949.447097] env[61970]: value = "task-1355913" [ 949.447097] env[61970]: _type = "Task" [ 949.447097] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.455017] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355913, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.463265] env[61970]: INFO nova.compute.manager [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Took 17.33 seconds to build instance. [ 949.657523] env[61970]: DEBUG oslo_vmware.api [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355910, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.629327} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.657783] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 949.657995] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 949.658212] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 949.658392] env[61970]: INFO nova.compute.manager [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Took 1.64 seconds to destroy the instance on the hypervisor. [ 949.658676] env[61970]: DEBUG oslo.service.loopingcall [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 949.658859] env[61970]: DEBUG nova.compute.manager [-] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 949.658960] env[61970]: DEBUG nova.network.neutron [-] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 949.729030] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 949.729285] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 949.729450] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 949.730052] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 949.730052] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 949.730052] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 949.730297] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 949.730297] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 949.730447] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 949.730616] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 949.730811] env[61970]: DEBUG nova.virt.hardware [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 949.738811] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Reconfiguring VM instance instance-00000054 to detach disk 2000 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 949.739372] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4245b633-17ef-4cb2-b14b-2bff67c6709b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.763376] env[61970]: DEBUG oslo_vmware.api [None req-b0f33a1f-83e8-493a-aaf9-e37fe62c6b84 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355912, 'name': PowerOffVM_Task, 'duration_secs': 0.352508} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.764566] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0f33a1f-83e8-493a-aaf9-e37fe62c6b84 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 949.764756] env[61970]: DEBUG nova.compute.manager [None req-b0f33a1f-83e8-493a-aaf9-e37fe62c6b84 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 949.765093] env[61970]: DEBUG oslo_vmware.api [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 949.765093] env[61970]: value = "task-1355914" [ 949.765093] env[61970]: _type = "Task" [ 949.765093] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.767824] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b07466-d23d-43f7-a440-5a1b6023bba0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.781249] env[61970]: DEBUG oslo_vmware.api [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355914, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.931284] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69944669-7016-4a93-b07d-89303615bfde {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.940193] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20f0fbac-dacb-4e21-b06b-8aa516c1ef3f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.959926] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355913, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.984165] env[61970]: DEBUG oslo_concurrency.lockutils [None req-135ae251-ee82-43e7-95ab-fa5ffab79ca4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "b9be8292-378b-4187-85dc-8d01e817faa3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.854s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.985467] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1fe1421-3fdb-469d-aed3-e1cdde436f2b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.994931] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e6bd125-fcab-4557-a0cf-574f53f7cb92 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.009788] env[61970]: DEBUG nova.compute.provider_tree [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.015040] env[61970]: DEBUG nova.compute.manager [req-850c4c48-5c47-4205-b2cd-d42a7bc26119 req-acd224f5-83b1-422f-b1ff-0175111942f5 service nova] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Received event network-vif-deleted-fa8c21c5-9351-49d0-8080-e693c96d1348 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 950.015396] env[61970]: INFO nova.compute.manager [req-850c4c48-5c47-4205-b2cd-d42a7bc26119 req-acd224f5-83b1-422f-b1ff-0175111942f5 service nova] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Neutron deleted interface fa8c21c5-9351-49d0-8080-e693c96d1348; detaching it from the instance and deleting it from the info cache [ 950.015590] env[61970]: DEBUG nova.network.neutron [req-850c4c48-5c47-4205-b2cd-d42a7bc26119 req-acd224f5-83b1-422f-b1ff-0175111942f5 service nova] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.025196] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ba2c8879-ec69-4c2a-a6da-a70515641b43 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "b9be8292-378b-4187-85dc-8d01e817faa3" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.025545] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ba2c8879-ec69-4c2a-a6da-a70515641b43 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "b9be8292-378b-4187-85dc-8d01e817faa3" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.025626] env[61970]: DEBUG nova.compute.manager [None req-ba2c8879-ec69-4c2a-a6da-a70515641b43 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 950.026995] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bc74ac6-03eb-4735-bb8f-cee22eb0f2b1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.034035] env[61970]: DEBUG nova.compute.manager [None req-ba2c8879-ec69-4c2a-a6da-a70515641b43 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61970) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 950.034623] env[61970]: DEBUG nova.objects.instance [None req-ba2c8879-ec69-4c2a-a6da-a70515641b43 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lazy-loading 'flavor' on Instance uuid b9be8292-378b-4187-85dc-8d01e817faa3 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.281071] env[61970]: DEBUG oslo_vmware.api [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355914, 'name': ReconfigVM_Task, 'duration_secs': 0.21187} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.281071] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Reconfigured VM instance instance-00000054 to detach disk 2000 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 950.281071] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f6859f-48f4-4cf5-bffe-0530be239032 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.285461] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0f33a1f-83e8-493a-aaf9-e37fe62c6b84 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.570s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.305812] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] dc38f9da-9f82-4707-9764-b28aa83a8ece/dc38f9da-9f82-4707-9764-b28aa83a8ece.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 950.306781] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-04eaf123-3cb4-4434-8a14-9096369e0790 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.327976] env[61970]: DEBUG oslo_vmware.api [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 950.327976] env[61970]: value = "task-1355915" [ 950.327976] env[61970]: _type = "Task" [ 950.327976] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.337256] env[61970]: DEBUG oslo_vmware.api [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355915, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.458333] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355913, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.788482} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.458603] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] fa8bf4b0-6955-4932-a85d-5a33947ae594/fa8bf4b0-6955-4932-a85d-5a33947ae594.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 950.458856] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 950.459130] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f861de33-f534-4271-b8ac-b1314decbda7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.465531] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 950.465531] env[61970]: value = "task-1355916" [ 950.465531] env[61970]: _type = "Task" [ 950.465531] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.473605] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355916, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.500274] env[61970]: DEBUG nova.network.neutron [-] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.512951] env[61970]: DEBUG nova.scheduler.client.report [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 950.517506] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c9f20601-56d5-4665-8c5a-35a69dfdec8b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.527857] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e54305b-6912-4822-b85a-35443d5ea4b6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.542886] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba2c8879-ec69-4c2a-a6da-a70515641b43 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 950.543119] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a46fdfea-0829-4e49-8447-2d84e682e010 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.551088] env[61970]: DEBUG oslo_vmware.api [None req-ba2c8879-ec69-4c2a-a6da-a70515641b43 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 950.551088] env[61970]: value = "task-1355917" [ 950.551088] env[61970]: _type = "Task" [ 950.551088] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.565505] env[61970]: DEBUG nova.compute.manager [req-850c4c48-5c47-4205-b2cd-d42a7bc26119 req-acd224f5-83b1-422f-b1ff-0175111942f5 service nova] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Detach interface failed, port_id=fa8c21c5-9351-49d0-8080-e693c96d1348, reason: Instance 0d21db81-a53f-4edf-a3a6-f1949cba2499 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 950.569450] env[61970]: DEBUG oslo_vmware.api [None req-ba2c8879-ec69-4c2a-a6da-a70515641b43 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355917, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.838361] env[61970]: DEBUG oslo_vmware.api [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355915, 'name': ReconfigVM_Task, 'duration_secs': 0.488931} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.838626] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Reconfigured VM instance instance-00000054 to attach disk [datastore1] dc38f9da-9f82-4707-9764-b28aa83a8ece/dc38f9da-9f82-4707-9764-b28aa83a8ece.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 950.838954] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Updating instance 'dc38f9da-9f82-4707-9764-b28aa83a8ece' progress to 50 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 950.976220] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355916, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099625} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.976372] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 950.977258] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f5f60a7-0eb0-472e-8708-c28c824a74db {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.000693] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] fa8bf4b0-6955-4932-a85d-5a33947ae594/fa8bf4b0-6955-4932-a85d-5a33947ae594.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 951.000862] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c10f300-a8af-4484-b523-b0580570f6e2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.014713] env[61970]: INFO nova.compute.manager [-] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Took 1.36 seconds to deallocate network for instance. [ 951.019659] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.330s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.020143] env[61970]: DEBUG nova.compute.manager [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 951.023625] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 951.023625] env[61970]: value = "task-1355918" [ 951.023625] env[61970]: _type = "Task" [ 951.023625] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.031778] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355918, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.060485] env[61970]: DEBUG oslo_vmware.api [None req-ba2c8879-ec69-4c2a-a6da-a70515641b43 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355917, 'name': PowerOffVM_Task, 'duration_secs': 0.227068} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.060773] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba2c8879-ec69-4c2a-a6da-a70515641b43 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 951.060955] env[61970]: DEBUG nova.compute.manager [None req-ba2c8879-ec69-4c2a-a6da-a70515641b43 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 951.061784] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f15e4b5b-1d7c-40cf-a896-35519c81788a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.180144] env[61970]: DEBUG nova.objects.instance [None req-c77d9848-1755-4cd7-9709-0f2b0b10ed15 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lazy-loading 'flavor' on Instance uuid d0ad9689-bdb7-4d68-b19a-3f805ab4612e {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.345995] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59a8c57f-298a-4d95-a88b-dab3dcf1783b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.365843] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a9e3002-cc82-43a4-8aa5-7736caef608b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.386448] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Updating instance 'dc38f9da-9f82-4707-9764-b28aa83a8ece' progress to 67 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 951.524761] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.524989] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.525234] env[61970]: DEBUG nova.objects.instance [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lazy-loading 'resources' on Instance uuid 0d21db81-a53f-4edf-a3a6-f1949cba2499 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.527115] env[61970]: DEBUG nova.compute.utils [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 951.528956] env[61970]: DEBUG nova.compute.manager [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 951.529135] env[61970]: DEBUG nova.network.neutron [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 951.540424] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355918, 'name': ReconfigVM_Task, 'duration_secs': 0.275494} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.540674] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Reconfigured VM instance instance-00000056 to attach disk [datastore2] fa8bf4b0-6955-4932-a85d-5a33947ae594/fa8bf4b0-6955-4932-a85d-5a33947ae594.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 951.541294] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-68afe7cb-0511-4c20-9d99-bd01159e5017 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.547308] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 951.547308] env[61970]: value = "task-1355919" [ 951.547308] env[61970]: _type = "Task" [ 951.547308] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.556797] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355919, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.572699] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ba2c8879-ec69-4c2a-a6da-a70515641b43 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "b9be8292-378b-4187-85dc-8d01e817faa3" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.547s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.577254] env[61970]: DEBUG nova.policy [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2af915dc43c84940a937dfb8a04ea0b6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17ca05a1e6664430a02de563d98c1148', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 951.684822] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c77d9848-1755-4cd7-9709-0f2b0b10ed15 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "refresh_cache-d0ad9689-bdb7-4d68-b19a-3f805ab4612e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.685023] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c77d9848-1755-4cd7-9709-0f2b0b10ed15 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquired lock "refresh_cache-d0ad9689-bdb7-4d68-b19a-3f805ab4612e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.685219] env[61970]: DEBUG nova.network.neutron [None req-c77d9848-1755-4cd7-9709-0f2b0b10ed15 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 951.685418] env[61970]: DEBUG nova.objects.instance [None req-c77d9848-1755-4cd7-9709-0f2b0b10ed15 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lazy-loading 'info_cache' on Instance uuid d0ad9689-bdb7-4d68-b19a-3f805ab4612e {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.894687] env[61970]: DEBUG nova.network.neutron [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Successfully created port: 4434695d-ddda-4840-b1da-8159c98dcae0 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 951.965882] env[61970]: DEBUG nova.network.neutron [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Port 313663a2-7d1f-40ea-abfc-c540d35235ca binding to destination host cpu-1 is already ACTIVE {{(pid=61970) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 952.035405] env[61970]: DEBUG nova.compute.manager [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 952.062030] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355919, 'name': Rename_Task, 'duration_secs': 0.140263} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.062030] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 952.062030] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-19f42c24-049c-4548-a943-ecde71b87c40 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.069575] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 952.069575] env[61970]: value = "task-1355920" [ 952.069575] env[61970]: _type = "Task" [ 952.069575] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.081981] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355920, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.190645] env[61970]: DEBUG nova.objects.base [None req-c77d9848-1755-4cd7-9709-0f2b0b10ed15 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 952.225786] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f5c1dec-0263-4511-b488-3897c28bc811 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.233823] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-356babde-dcf9-4036-a03f-a019ce9a7edb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.268378] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42f0030-39d9-48c3-ad70-f3a64dc34b6c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.276206] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e7c47d-03cb-4cef-83f6-fbf025eac613 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.289194] env[61970]: DEBUG nova.compute.provider_tree [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.475866] env[61970]: DEBUG oslo_concurrency.lockutils [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "b9be8292-378b-4187-85dc-8d01e817faa3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.476232] env[61970]: DEBUG oslo_concurrency.lockutils [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "b9be8292-378b-4187-85dc-8d01e817faa3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.476336] env[61970]: DEBUG oslo_concurrency.lockutils [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "b9be8292-378b-4187-85dc-8d01e817faa3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.476809] env[61970]: DEBUG oslo_concurrency.lockutils [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "b9be8292-378b-4187-85dc-8d01e817faa3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.476809] env[61970]: DEBUG oslo_concurrency.lockutils [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "b9be8292-378b-4187-85dc-8d01e817faa3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.478885] env[61970]: INFO nova.compute.manager [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Terminating instance [ 952.480789] env[61970]: DEBUG nova.compute.manager [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 952.480987] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 952.481852] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71633be3-b10f-4d9b-a171-a174758f6a72 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.489577] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 952.489862] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6f30ec93-2c96-4c99-ac34-0339df3f66b6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.550055] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 952.550290] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 952.550439] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Deleting the datastore file [datastore1] b9be8292-378b-4187-85dc-8d01e817faa3 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 952.550696] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5e726cd2-c030-413c-8690-3a520c42eb0b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.556920] env[61970]: DEBUG oslo_vmware.api [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 952.556920] env[61970]: value = "task-1355922" [ 952.556920] env[61970]: _type = "Task" [ 952.556920] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.565788] env[61970]: DEBUG oslo_vmware.api [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355922, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.580588] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355920, 'name': PowerOnVM_Task, 'duration_secs': 0.439094} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.580903] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 952.581231] env[61970]: DEBUG nova.compute.manager [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 952.582124] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76c6898c-5e81-4d4a-ab72-24d56f796474 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.791839] env[61970]: DEBUG nova.scheduler.client.report [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.901590] env[61970]: DEBUG nova.network.neutron [None req-c77d9848-1755-4cd7-9709-0f2b0b10ed15 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Updating instance_info_cache with network_info: [{"id": "b83024be-4171-40cd-9d1c-65a2d2ce8b9e", "address": "fa:16:3e:f5:18:c9", "network": {"id": "1f53a96b-8a46-41d5-8bf3-759de1fce443", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-108385347-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a44c05a894394d69b8b4fd36c7b2532a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb83024be-41", "ovs_interfaceid": "b83024be-4171-40cd-9d1c-65a2d2ce8b9e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.986035] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "dc38f9da-9f82-4707-9764-b28aa83a8ece-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.986035] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "dc38f9da-9f82-4707-9764-b28aa83a8ece-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.986224] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "dc38f9da-9f82-4707-9764-b28aa83a8ece-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.045948] env[61970]: DEBUG nova.compute.manager [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 953.069624] env[61970]: DEBUG oslo_vmware.api [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355922, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14235} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.069975] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 953.070078] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 953.070249] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 953.070417] env[61970]: INFO nova.compute.manager [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Took 0.59 seconds to destroy the instance on the hypervisor. [ 953.070660] env[61970]: DEBUG oslo.service.loopingcall [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.070852] env[61970]: DEBUG nova.compute.manager [-] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 953.070946] env[61970]: DEBUG nova.network.neutron [-] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 953.075012] env[61970]: DEBUG nova.virt.hardware [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 953.075259] env[61970]: DEBUG nova.virt.hardware [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 953.075423] env[61970]: DEBUG nova.virt.hardware [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 953.075606] env[61970]: DEBUG nova.virt.hardware [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 953.075755] env[61970]: DEBUG nova.virt.hardware [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 953.075904] env[61970]: DEBUG nova.virt.hardware [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 953.076130] env[61970]: DEBUG nova.virt.hardware [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 953.076299] env[61970]: DEBUG nova.virt.hardware [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 953.076594] env[61970]: DEBUG nova.virt.hardware [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 953.076945] env[61970]: DEBUG nova.virt.hardware [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 953.077169] env[61970]: DEBUG nova.virt.hardware [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 953.077928] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-019eaa1f-3dbe-4919-87dd-e9f1aabdc51e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.085562] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7496565f-bd22-4765-828e-b43c8eadc2f4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.100540] env[61970]: INFO nova.compute.manager [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] bringing vm to original state: 'stopped' [ 953.298050] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.773s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.322582] env[61970]: INFO nova.scheduler.client.report [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleted allocations for instance 0d21db81-a53f-4edf-a3a6-f1949cba2499 [ 953.406981] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c77d9848-1755-4cd7-9709-0f2b0b10ed15 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Releasing lock "refresh_cache-d0ad9689-bdb7-4d68-b19a-3f805ab4612e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.429069] env[61970]: DEBUG nova.compute.manager [req-019990ee-336f-417d-aada-e5f6a9d640d2 req-1e3ad32b-5961-4956-a48d-cc2fc084e92d service nova] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Received event network-vif-deleted-490f6a87-12ad-4d7f-9dc5-46d694adf506 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.429316] env[61970]: INFO nova.compute.manager [req-019990ee-336f-417d-aada-e5f6a9d640d2 req-1e3ad32b-5961-4956-a48d-cc2fc084e92d service nova] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Neutron deleted interface 490f6a87-12ad-4d7f-9dc5-46d694adf506; detaching it from the instance and deleting it from the info cache [ 953.429496] env[61970]: DEBUG nova.network.neutron [req-019990ee-336f-417d-aada-e5f6a9d640d2 req-1e3ad32b-5961-4956-a48d-cc2fc084e92d service nova] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.659175] env[61970]: DEBUG nova.network.neutron [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Successfully updated port: 4434695d-ddda-4840-b1da-8159c98dcae0 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 953.830204] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6a2f2a14-cd0d-4eb4-a009-b643f6a7b8b4 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "0d21db81-a53f-4edf-a3a6-f1949cba2499" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.823s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.891212] env[61970]: DEBUG nova.network.neutron [-] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.910558] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c77d9848-1755-4cd7-9709-0f2b0b10ed15 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 953.910846] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e16b887d-47b7-42e3-9fc3-a5cdb91f7ef4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.918411] env[61970]: DEBUG oslo_vmware.api [None req-c77d9848-1755-4cd7-9709-0f2b0b10ed15 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 953.918411] env[61970]: value = "task-1355923" [ 953.918411] env[61970]: _type = "Task" [ 953.918411] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.926422] env[61970]: DEBUG oslo_vmware.api [None req-c77d9848-1755-4cd7-9709-0f2b0b10ed15 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355923, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.934894] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb38e1bf-07c7-4ea3-a7c7-81d748f2fe55 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.943963] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba98594-2f33-47b6-8801-dc7335b6f837 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.973993] env[61970]: DEBUG nova.compute.manager [req-019990ee-336f-417d-aada-e5f6a9d640d2 req-1e3ad32b-5961-4956-a48d-cc2fc084e92d service nova] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Detach interface failed, port_id=490f6a87-12ad-4d7f-9dc5-46d694adf506, reason: Instance b9be8292-378b-4187-85dc-8d01e817faa3 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 954.043184] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "refresh_cache-dc38f9da-9f82-4707-9764-b28aa83a8ece" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.043378] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "refresh_cache-dc38f9da-9f82-4707-9764-b28aa83a8ece" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.043555] env[61970]: DEBUG nova.network.neutron [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 954.109667] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "fa8bf4b0-6955-4932-a85d-5a33947ae594" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.110035] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "fa8bf4b0-6955-4932-a85d-5a33947ae594" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.110246] env[61970]: DEBUG nova.compute.manager [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.111149] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5d802c-aa73-4f5f-8f6e-2a3d6bc99d79 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.118567] env[61970]: DEBUG nova.compute.manager [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61970) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 954.120777] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.121036] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5b6fbc0-d406-4ccf-9412-41d2fafffacc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.127654] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 954.127654] env[61970]: value = "task-1355924" [ 954.127654] env[61970]: _type = "Task" [ 954.127654] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.135561] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355924, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.160676] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.160856] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.161010] env[61970]: DEBUG nova.network.neutron [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 954.393910] env[61970]: INFO nova.compute.manager [-] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Took 1.32 seconds to deallocate network for instance. [ 954.429585] env[61970]: DEBUG oslo_vmware.api [None req-c77d9848-1755-4cd7-9709-0f2b0b10ed15 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1355923, 'name': PowerOnVM_Task, 'duration_secs': 0.404057} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.429912] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c77d9848-1755-4cd7-9709-0f2b0b10ed15 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 954.430178] env[61970]: DEBUG nova.compute.manager [None req-c77d9848-1755-4cd7-9709-0f2b0b10ed15 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.431320] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10456657-bc3b-492b-9665-852070695385 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.638857] env[61970]: DEBUG oslo_vmware.api [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355924, 'name': PowerOffVM_Task, 'duration_secs': 0.226637} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.638857] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.638857] env[61970]: DEBUG nova.compute.manager [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.639848] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e07d5a63-dc02-4551-bd06-c5746e4b9883 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.642574] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "3b346c68-53a9-4ceb-83a3-9e4bce9610b1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.642845] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "3b346c68-53a9-4ceb-83a3-9e4bce9610b1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.643070] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "3b346c68-53a9-4ceb-83a3-9e4bce9610b1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.643273] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "3b346c68-53a9-4ceb-83a3-9e4bce9610b1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.643443] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "3b346c68-53a9-4ceb-83a3-9e4bce9610b1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.645608] env[61970]: INFO nova.compute.manager [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Terminating instance [ 954.652905] env[61970]: DEBUG nova.compute.manager [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 954.653111] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 954.654144] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab6c9a7-d66a-419f-9b00-40aae15e8bbc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.661079] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.661391] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed77044c-3b88-455e-819d-c7ac19c275c4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.669189] env[61970]: DEBUG oslo_vmware.api [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 954.669189] env[61970]: value = "task-1355925" [ 954.669189] env[61970]: _type = "Task" [ 954.669189] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.676551] env[61970]: DEBUG oslo_vmware.api [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355925, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.731285] env[61970]: DEBUG nova.network.neutron [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 954.879292] env[61970]: DEBUG nova.network.neutron [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Updating instance_info_cache with network_info: [{"id": "313663a2-7d1f-40ea-abfc-c540d35235ca", "address": "fa:16:3e:c7:5e:f0", "network": {"id": "cba185ff-bfa3-4547-b7b0-638c3e79a24c", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-864852491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5c64cb9232fb413cbd7627dcf077e9ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap313663a2-7d", "ovs_interfaceid": "313663a2-7d1f-40ea-abfc-c540d35235ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.904317] env[61970]: DEBUG oslo_concurrency.lockutils [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.904840] env[61970]: DEBUG oslo_concurrency.lockutils [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.905022] env[61970]: DEBUG nova.objects.instance [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lazy-loading 'resources' on Instance uuid b9be8292-378b-4187-85dc-8d01e817faa3 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.941319] env[61970]: DEBUG nova.network.neutron [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Updating instance_info_cache with network_info: [{"id": "4434695d-ddda-4840-b1da-8159c98dcae0", "address": "fa:16:3e:85:27:97", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4434695d-dd", "ovs_interfaceid": "4434695d-ddda-4840-b1da-8159c98dcae0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.161119] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "fa8bf4b0-6955-4932-a85d-5a33947ae594" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.050s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.179446] env[61970]: DEBUG oslo_vmware.api [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355925, 'name': PowerOffVM_Task, 'duration_secs': 0.245927} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.179708] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.179883] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.180152] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b84b0288-97b7-48e1-ae2d-34b424ab7a88 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.248922] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 955.249189] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 955.249379] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleting the datastore file [datastore2] 3b346c68-53a9-4ceb-83a3-9e4bce9610b1 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.249846] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e337c8c-e3cf-4ead-a891-86e721667d02 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.256511] env[61970]: DEBUG oslo_vmware.api [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 955.256511] env[61970]: value = "task-1355927" [ 955.256511] env[61970]: _type = "Task" [ 955.256511] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.265680] env[61970]: DEBUG oslo_vmware.api [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355927, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.382568] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "refresh_cache-dc38f9da-9f82-4707-9764-b28aa83a8ece" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.445428] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.445814] env[61970]: DEBUG nova.compute.manager [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Instance network_info: |[{"id": "4434695d-ddda-4840-b1da-8159c98dcae0", "address": "fa:16:3e:85:27:97", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4434695d-dd", "ovs_interfaceid": "4434695d-ddda-4840-b1da-8159c98dcae0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 955.446441] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:27:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '68add7d6-c025-46fa-84d3-9c589adb63e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4434695d-ddda-4840-b1da-8159c98dcae0', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 955.453849] env[61970]: DEBUG oslo.service.loopingcall [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.457470] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 955.459011] env[61970]: DEBUG nova.compute.manager [req-085c9dd6-d567-427c-bf3e-9e1d4934f7cf req-45860c47-b12e-4738-bc40-05d7ef2750b1 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Received event network-vif-plugged-4434695d-ddda-4840-b1da-8159c98dcae0 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.459250] env[61970]: DEBUG oslo_concurrency.lockutils [req-085c9dd6-d567-427c-bf3e-9e1d4934f7cf req-45860c47-b12e-4738-bc40-05d7ef2750b1 service nova] Acquiring lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.459458] env[61970]: DEBUG oslo_concurrency.lockutils [req-085c9dd6-d567-427c-bf3e-9e1d4934f7cf req-45860c47-b12e-4738-bc40-05d7ef2750b1 service nova] Lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.459627] env[61970]: DEBUG oslo_concurrency.lockutils [req-085c9dd6-d567-427c-bf3e-9e1d4934f7cf req-45860c47-b12e-4738-bc40-05d7ef2750b1 service nova] Lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.459795] env[61970]: DEBUG nova.compute.manager [req-085c9dd6-d567-427c-bf3e-9e1d4934f7cf req-45860c47-b12e-4738-bc40-05d7ef2750b1 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] No waiting events found dispatching network-vif-plugged-4434695d-ddda-4840-b1da-8159c98dcae0 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 955.460010] env[61970]: WARNING nova.compute.manager [req-085c9dd6-d567-427c-bf3e-9e1d4934f7cf req-45860c47-b12e-4738-bc40-05d7ef2750b1 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Received unexpected event network-vif-plugged-4434695d-ddda-4840-b1da-8159c98dcae0 for instance with vm_state building and task_state spawning. [ 955.460191] env[61970]: DEBUG nova.compute.manager [req-085c9dd6-d567-427c-bf3e-9e1d4934f7cf req-45860c47-b12e-4738-bc40-05d7ef2750b1 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Received event network-changed-4434695d-ddda-4840-b1da-8159c98dcae0 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.460353] env[61970]: DEBUG nova.compute.manager [req-085c9dd6-d567-427c-bf3e-9e1d4934f7cf req-45860c47-b12e-4738-bc40-05d7ef2750b1 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Refreshing instance network info cache due to event network-changed-4434695d-ddda-4840-b1da-8159c98dcae0. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 955.460540] env[61970]: DEBUG oslo_concurrency.lockutils [req-085c9dd6-d567-427c-bf3e-9e1d4934f7cf req-45860c47-b12e-4738-bc40-05d7ef2750b1 service nova] Acquiring lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.460748] env[61970]: DEBUG oslo_concurrency.lockutils [req-085c9dd6-d567-427c-bf3e-9e1d4934f7cf req-45860c47-b12e-4738-bc40-05d7ef2750b1 service nova] Acquired lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.460928] env[61970]: DEBUG nova.network.neutron [req-085c9dd6-d567-427c-bf3e-9e1d4934f7cf req-45860c47-b12e-4738-bc40-05d7ef2750b1 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Refreshing network info cache for port 4434695d-ddda-4840-b1da-8159c98dcae0 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 955.462880] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c459a0d7-a03b-4a72-9a80-88198fba821a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.485878] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 955.485878] env[61970]: value = "task-1355928" [ 955.485878] env[61970]: _type = "Task" [ 955.485878] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.494855] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355928, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.600050] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d2d413-37d9-46f6-aa3f-d5792b2fc904 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.607620] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d881ebfc-07fb-4e7e-ac9d-d960a70297d5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.638050] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dd8a15e-8387-4f89-adae-dfef1618a457 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.645459] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b9ab832-82ac-4130-891b-983499d0a83e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.659032] env[61970]: DEBUG nova.compute.provider_tree [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.669836] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.750377] env[61970]: DEBUG oslo_concurrency.lockutils [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "7f59000a-94d9-45b6-aa7a-300d95793615" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.750654] env[61970]: DEBUG oslo_concurrency.lockutils [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "7f59000a-94d9-45b6-aa7a-300d95793615" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.750845] env[61970]: INFO nova.compute.manager [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Shelving [ 955.768346] env[61970]: DEBUG oslo_vmware.api [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355927, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.125108} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.768592] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 955.768784] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 955.769074] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 955.769429] env[61970]: INFO nova.compute.manager [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Took 1.12 seconds to destroy the instance on the hypervisor. [ 955.769791] env[61970]: DEBUG oslo.service.loopingcall [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.770036] env[61970]: DEBUG nova.compute.manager [-] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 955.770147] env[61970]: DEBUG nova.network.neutron [-] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 955.908162] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f25c608-df53-4ad1-8838-2a41da9a48a4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.930027] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2447920-09ff-4e11-8a98-788627ffc9a6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.937092] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Updating instance 'dc38f9da-9f82-4707-9764-b28aa83a8ece' progress to 83 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 956.000021] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355928, 'name': CreateVM_Task, 'duration_secs': 0.406511} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.000021] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 956.000021] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.000021] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.000021] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 956.000021] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34d90b73-0cde-4b53-9c2b-0d8c4e46f62e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.004996] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 956.004996] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b7cb61-35c5-4004-1940-a6cbe0c69ba3" [ 956.004996] env[61970]: _type = "Task" [ 956.004996] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.013867] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b7cb61-35c5-4004-1940-a6cbe0c69ba3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.162188] env[61970]: DEBUG nova.scheduler.client.report [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.224690] env[61970]: DEBUG nova.network.neutron [req-085c9dd6-d567-427c-bf3e-9e1d4934f7cf req-45860c47-b12e-4738-bc40-05d7ef2750b1 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Updated VIF entry in instance network info cache for port 4434695d-ddda-4840-b1da-8159c98dcae0. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 956.225063] env[61970]: DEBUG nova.network.neutron [req-085c9dd6-d567-427c-bf3e-9e1d4934f7cf req-45860c47-b12e-4738-bc40-05d7ef2750b1 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Updating instance_info_cache with network_info: [{"id": "4434695d-ddda-4840-b1da-8159c98dcae0", "address": "fa:16:3e:85:27:97", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4434695d-dd", "ovs_interfaceid": "4434695d-ddda-4840-b1da-8159c98dcae0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.258489] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 956.258743] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5ab0d6c-f83e-4a64-a721-2429e94d9f19 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.266394] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 956.266394] env[61970]: value = "task-1355929" [ 956.266394] env[61970]: _type = "Task" [ 956.266394] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.275536] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355929, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.443008] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 956.443316] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a7aea92-865e-46c0-bb67-0d24a7eda486 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.451263] env[61970]: DEBUG oslo_vmware.api [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 956.451263] env[61970]: value = "task-1355930" [ 956.451263] env[61970]: _type = "Task" [ 956.451263] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.459677] env[61970]: DEBUG oslo_vmware.api [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355930, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.517573] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b7cb61-35c5-4004-1940-a6cbe0c69ba3, 'name': SearchDatastore_Task, 'duration_secs': 0.010354} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.517950] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.518247] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 956.518550] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.518711] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.518899] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 956.519267] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ba50fcc6-2ba6-4991-b30c-09c70c0e2918 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.549261] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "fa8bf4b0-6955-4932-a85d-5a33947ae594" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.549561] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "fa8bf4b0-6955-4932-a85d-5a33947ae594" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.549786] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "fa8bf4b0-6955-4932-a85d-5a33947ae594-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.550037] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "fa8bf4b0-6955-4932-a85d-5a33947ae594-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.550221] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "fa8bf4b0-6955-4932-a85d-5a33947ae594-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.552540] env[61970]: INFO nova.compute.manager [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Terminating instance [ 956.554310] env[61970]: DEBUG nova.compute.manager [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 956.554502] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 956.555433] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192c9dbb-81d6-4d4d-87b0-fb82e300306e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.560359] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 956.560534] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 956.563723] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7c8e674-b090-4374-b5cf-7eabe9cff837 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.566407] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 956.566667] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-155200e4-e1c0-4407-a248-45185c513f79 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.571016] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 956.571016] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5264d48b-390f-7565-d616-05b07c067064" [ 956.571016] env[61970]: _type = "Task" [ 956.571016] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.578633] env[61970]: DEBUG nova.network.neutron [-] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.579883] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5264d48b-390f-7565-d616-05b07c067064, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.629305] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 956.629583] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 956.629838] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleting the datastore file [datastore2] fa8bf4b0-6955-4932-a85d-5a33947ae594 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.630216] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9d21544d-fc19-4484-8e75-7d0c59a97107 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.636602] env[61970]: DEBUG oslo_vmware.api [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 956.636602] env[61970]: value = "task-1355932" [ 956.636602] env[61970]: _type = "Task" [ 956.636602] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.646214] env[61970]: DEBUG oslo_vmware.api [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355932, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.666757] env[61970]: DEBUG oslo_concurrency.lockutils [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.762s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.669181] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 1.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.669411] env[61970]: DEBUG nova.objects.instance [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61970) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 956.693367] env[61970]: INFO nova.scheduler.client.report [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Deleted allocations for instance b9be8292-378b-4187-85dc-8d01e817faa3 [ 956.727823] env[61970]: DEBUG oslo_concurrency.lockutils [req-085c9dd6-d567-427c-bf3e-9e1d4934f7cf req-45860c47-b12e-4738-bc40-05d7ef2750b1 service nova] Releasing lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.777082] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355929, 'name': PowerOffVM_Task, 'duration_secs': 0.221495} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.777082] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 956.777398] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd7fcd8-5496-4524-9c03-df44360c4870 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.795147] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a897084-44fe-4e83-8314-c2dc03499ac2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.963293] env[61970]: DEBUG oslo_vmware.api [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355930, 'name': PowerOnVM_Task, 'duration_secs': 0.461457} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.964047] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 956.964047] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d8bf3c22-1506-4f82-bbcb-cd1e3fc55114 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Updating instance 'dc38f9da-9f82-4707-9764-b28aa83a8ece' progress to 100 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 957.081624] env[61970]: INFO nova.compute.manager [-] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Took 1.31 seconds to deallocate network for instance. [ 957.081964] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5264d48b-390f-7565-d616-05b07c067064, 'name': SearchDatastore_Task, 'duration_secs': 0.427697} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.086351] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c7b5bd8-9024-4b17-a2f9-91227eacd217 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.092126] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 957.092126] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525bfafc-0057-301a-8ab9-5b3bac6c4fee" [ 957.092126] env[61970]: _type = "Task" [ 957.092126] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.099983] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525bfafc-0057-301a-8ab9-5b3bac6c4fee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.146383] env[61970]: DEBUG oslo_vmware.api [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355932, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137651} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.146618] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 957.146802] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 957.146983] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 957.147172] env[61970]: INFO nova.compute.manager [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Took 0.59 seconds to destroy the instance on the hypervisor. [ 957.147409] env[61970]: DEBUG oslo.service.loopingcall [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.147597] env[61970]: DEBUG nova.compute.manager [-] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 957.147691] env[61970]: DEBUG nova.network.neutron [-] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 957.205308] env[61970]: DEBUG oslo_concurrency.lockutils [None req-372e18c4-d0b1-43ab-a232-ea207de90dbf tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "b9be8292-378b-4187-85dc-8d01e817faa3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.729s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.305990] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Creating Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 957.306384] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-7855212e-6a3e-412c-826e-435e4fe6d74b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.314611] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 957.314611] env[61970]: value = "task-1355933" [ 957.314611] env[61970]: _type = "Task" [ 957.314611] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.323099] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355933, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.493742] env[61970]: DEBUG nova.compute.manager [req-87131252-359f-4ce1-91da-4164733a1e5d req-6f76ce95-2427-49e3-af40-34bccb9b0a4b service nova] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Received event network-vif-deleted-10059ea8-815e-48c8-9dcb-8da7857f570a {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 957.493903] env[61970]: DEBUG nova.compute.manager [req-87131252-359f-4ce1-91da-4164733a1e5d req-6f76ce95-2427-49e3-af40-34bccb9b0a4b service nova] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Received event network-vif-deleted-df21eb74-2d16-49a2-9a94-372df5b2f8a7 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 957.494096] env[61970]: INFO nova.compute.manager [req-87131252-359f-4ce1-91da-4164733a1e5d req-6f76ce95-2427-49e3-af40-34bccb9b0a4b service nova] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Neutron deleted interface df21eb74-2d16-49a2-9a94-372df5b2f8a7; detaching it from the instance and deleting it from the info cache [ 957.494269] env[61970]: DEBUG nova.network.neutron [req-87131252-359f-4ce1-91da-4164733a1e5d req-6f76ce95-2427-49e3-af40-34bccb9b0a4b service nova] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.590227] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.602147] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525bfafc-0057-301a-8ab9-5b3bac6c4fee, 'name': SearchDatastore_Task, 'duration_secs': 0.123417} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.602412] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.602670] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8/c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 957.602938] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fbd7b6fc-8b5a-4c5e-89b5-35592ea3064f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.609789] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 957.609789] env[61970]: value = "task-1355934" [ 957.609789] env[61970]: _type = "Task" [ 957.609789] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.617852] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355934, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.679303] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60d635c7-8710-428a-baaf-58cf6b237a6f tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.680528] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.090s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.680665] env[61970]: DEBUG nova.objects.instance [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lazy-loading 'resources' on Instance uuid 3b346c68-53a9-4ceb-83a3-9e4bce9610b1 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.824632] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355933, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.885254] env[61970]: DEBUG nova.network.neutron [-] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.997727] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2c06c3e8-2d4e-478d-9a44-d6a981cc0fc9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.009118] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3e6ce1a-2e36-48b7-a736-8418d01d82d7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.052297] env[61970]: DEBUG nova.compute.manager [req-87131252-359f-4ce1-91da-4164733a1e5d req-6f76ce95-2427-49e3-af40-34bccb9b0a4b service nova] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Detach interface failed, port_id=df21eb74-2d16-49a2-9a94-372df5b2f8a7, reason: Instance fa8bf4b0-6955-4932-a85d-5a33947ae594 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 958.122048] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355934, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.325377] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355933, 'name': CreateSnapshot_Task, 'duration_secs': 0.877146} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.325652] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Created Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 958.326402] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89cabaf2-cef5-4819-abbc-4bad8e9cc206 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.358503] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aba5273-7489-4079-9297-fe5e3316e209 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.366620] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd0d0749-fb33-4f6e-8462-47e1e38b324e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.396449] env[61970]: INFO nova.compute.manager [-] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Took 1.25 seconds to deallocate network for instance. [ 958.398638] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55323487-4e0d-4929-b6db-f3d749481af3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.410345] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d5232f-0654-4ab1-8359-50f7b0c7a749 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.431783] env[61970]: DEBUG nova.compute.provider_tree [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.578526] env[61970]: DEBUG oslo_concurrency.lockutils [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "479342e0-7627-4dda-a589-a7bc5a5c4270" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.578766] env[61970]: DEBUG oslo_concurrency.lockutils [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "479342e0-7627-4dda-a589-a7bc5a5c4270" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.621007] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355934, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.620388} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.621277] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8/c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 958.621501] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 958.622021] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-108b0111-486b-416b-994e-dd6bbb40029c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.628257] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 958.628257] env[61970]: value = "task-1355935" [ 958.628257] env[61970]: _type = "Task" [ 958.628257] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.635147] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355935, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.846382] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Creating linked-clone VM from snapshot {{(pid=61970) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 958.847088] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5dcf13eb-edd7-461c-a270-61149db5b1dd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.855589] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 958.855589] env[61970]: value = "task-1355936" [ 958.855589] env[61970]: _type = "Task" [ 958.855589] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.863326] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355936, 'name': CloneVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.905535] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.936022] env[61970]: DEBUG nova.scheduler.client.report [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 959.080626] env[61970]: DEBUG nova.compute.manager [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 959.137828] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355935, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062699} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.138141] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 959.138876] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2141b011-ffea-4adc-89d1-6a82c9501630 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.161841] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8/c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 959.162122] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2880d146-dc38-4786-b8bd-78fb3af64621 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.181459] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 959.181459] env[61970]: value = "task-1355937" [ 959.181459] env[61970]: _type = "Task" [ 959.181459] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.189388] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355937, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.366069] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355936, 'name': CloneVM_Task} progress is 94%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.440451] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.760s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.443050] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.537s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.443183] env[61970]: DEBUG nova.objects.instance [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lazy-loading 'resources' on Instance uuid fa8bf4b0-6955-4932-a85d-5a33947ae594 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.464932] env[61970]: INFO nova.scheduler.client.report [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleted allocations for instance 3b346c68-53a9-4ceb-83a3-9e4bce9610b1 [ 959.487278] env[61970]: DEBUG oslo_concurrency.lockutils [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "dc38f9da-9f82-4707-9764-b28aa83a8ece" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.487708] env[61970]: DEBUG oslo_concurrency.lockutils [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "dc38f9da-9f82-4707-9764-b28aa83a8ece" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.488025] env[61970]: DEBUG nova.compute.manager [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Going to confirm migration 2 {{(pid=61970) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 959.600249] env[61970]: DEBUG oslo_concurrency.lockutils [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.691638] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355937, 'name': ReconfigVM_Task, 'duration_secs': 0.304495} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.692069] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Reconfigured VM instance instance-0000005a to attach disk [datastore1] c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8/c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 959.692819] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4d6297be-b072-48d0-9f4f-55cb559e8001 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.699382] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 959.699382] env[61970]: value = "task-1355938" [ 959.699382] env[61970]: _type = "Task" [ 959.699382] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.707463] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355938, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.865588] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355936, 'name': CloneVM_Task} progress is 95%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.974427] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f26f7212-87b2-464b-b363-d5a8dcc60d47 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "3b346c68-53a9-4ceb-83a3-9e4bce9610b1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.331s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.058393] env[61970]: DEBUG oslo_concurrency.lockutils [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "refresh_cache-dc38f9da-9f82-4707-9764-b28aa83a8ece" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.058580] env[61970]: DEBUG oslo_concurrency.lockutils [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "refresh_cache-dc38f9da-9f82-4707-9764-b28aa83a8ece" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.058756] env[61970]: DEBUG nova.network.neutron [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 960.058944] env[61970]: DEBUG nova.objects.instance [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lazy-loading 'info_cache' on Instance uuid dc38f9da-9f82-4707-9764-b28aa83a8ece {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.102885] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7ecfa87-d06f-423e-be5b-ef0eda0337ed {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.110808] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8071b60-d79c-40ba-976a-2e80a806161d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.144498] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5186bf1a-d117-422a-aa5a-fb90f2f4ca19 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.152714] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d675c985-1588-45eb-8078-ce07ecf4896f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.167637] env[61970]: DEBUG nova.compute.provider_tree [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 960.210114] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355938, 'name': Rename_Task, 'duration_secs': 0.151932} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.210114] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 960.210338] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-00f014d4-cd5a-49f1-af2c-a06f878776ec {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.216542] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 960.216542] env[61970]: value = "task-1355939" [ 960.216542] env[61970]: _type = "Task" [ 960.216542] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.224215] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355939, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.367743] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355936, 'name': CloneVM_Task, 'duration_secs': 1.321021} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.368017] env[61970]: INFO nova.virt.vmwareapi.vmops [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Created linked-clone VM from snapshot [ 960.368901] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ad2a06-7925-4757-8189-5de89b3b9878 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.376236] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Uploading image 8ff34e61-3203-40b2-805a-fd5355a36d65 {{(pid=61970) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 960.404797] env[61970]: DEBUG oslo_vmware.rw_handles [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 960.404797] env[61970]: value = "vm-288846" [ 960.404797] env[61970]: _type = "VirtualMachine" [ 960.404797] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 960.405104] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-153afdcb-d9a7-4221-b78b-a74401f17459 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.412152] env[61970]: DEBUG oslo_vmware.rw_handles [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lease: (returnval){ [ 960.412152] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5268e3ed-7b20-f10e-6eda-3fea15243003" [ 960.412152] env[61970]: _type = "HttpNfcLease" [ 960.412152] env[61970]: } obtained for exporting VM: (result){ [ 960.412152] env[61970]: value = "vm-288846" [ 960.412152] env[61970]: _type = "VirtualMachine" [ 960.412152] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 960.412451] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the lease: (returnval){ [ 960.412451] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5268e3ed-7b20-f10e-6eda-3fea15243003" [ 960.412451] env[61970]: _type = "HttpNfcLease" [ 960.412451] env[61970]: } to be ready. {{(pid=61970) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 960.419531] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 960.419531] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5268e3ed-7b20-f10e-6eda-3fea15243003" [ 960.419531] env[61970]: _type = "HttpNfcLease" [ 960.419531] env[61970]: } is initializing. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 960.671015] env[61970]: DEBUG nova.scheduler.client.report [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.726639] env[61970]: DEBUG oslo_vmware.api [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355939, 'name': PowerOnVM_Task, 'duration_secs': 0.461287} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.726921] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 960.727148] env[61970]: INFO nova.compute.manager [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Took 7.68 seconds to spawn the instance on the hypervisor. [ 960.727364] env[61970]: DEBUG nova.compute.manager [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 960.728170] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23c624f-00de-4020-9643-86362c93fe78 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.782291] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "4d093692-07d3-49aa-8b0a-fb52bd9c0ffc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.782518] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "4d093692-07d3-49aa-8b0a-fb52bd9c0ffc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.922378] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 960.922378] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5268e3ed-7b20-f10e-6eda-3fea15243003" [ 960.922378] env[61970]: _type = "HttpNfcLease" [ 960.922378] env[61970]: } is ready. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 960.922734] env[61970]: DEBUG oslo_vmware.rw_handles [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 960.922734] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5268e3ed-7b20-f10e-6eda-3fea15243003" [ 960.922734] env[61970]: _type = "HttpNfcLease" [ 960.922734] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 960.923424] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8ad599-f065-43f8-a5ae-99043793488f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.931474] env[61970]: DEBUG oslo_vmware.rw_handles [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526053fe-3aff-2103-cc95-816f32571d5a/disk-0.vmdk from lease info. {{(pid=61970) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 960.931659] env[61970]: DEBUG oslo_vmware.rw_handles [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526053fe-3aff-2103-cc95-816f32571d5a/disk-0.vmdk for reading. {{(pid=61970) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 961.023346] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d7e90d26-11ff-43f8-8375-7ab4fa31798a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.176065] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.733s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.178012] env[61970]: DEBUG oslo_concurrency.lockutils [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.578s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.179612] env[61970]: INFO nova.compute.claims [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 961.201684] env[61970]: INFO nova.scheduler.client.report [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleted allocations for instance fa8bf4b0-6955-4932-a85d-5a33947ae594 [ 961.247622] env[61970]: INFO nova.compute.manager [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Took 12.58 seconds to build instance. [ 961.285460] env[61970]: DEBUG nova.compute.manager [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 961.306085] env[61970]: DEBUG nova.network.neutron [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Updating instance_info_cache with network_info: [{"id": "313663a2-7d1f-40ea-abfc-c540d35235ca", "address": "fa:16:3e:c7:5e:f0", "network": {"id": "cba185ff-bfa3-4547-b7b0-638c3e79a24c", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-864852491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5c64cb9232fb413cbd7627dcf077e9ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap313663a2-7d", "ovs_interfaceid": "313663a2-7d1f-40ea-abfc-c540d35235ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.605549] env[61970]: DEBUG nova.compute.manager [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 961.606712] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbdfdd47-e1e3-4451-a697-2944810330c3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.710600] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d6c6bf61-22dc-46b4-b8e5-e5f6e8b04770 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "fa8bf4b0-6955-4932-a85d-5a33947ae594" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.161s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.749548] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b435c594-eed0-4d8e-a40b-8f1b5e85784e tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.096s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.809026] env[61970]: DEBUG oslo_concurrency.lockutils [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "refresh_cache-dc38f9da-9f82-4707-9764-b28aa83a8ece" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.809523] env[61970]: DEBUG nova.objects.instance [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lazy-loading 'migration_context' on Instance uuid dc38f9da-9f82-4707-9764-b28aa83a8ece {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.812020] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.118927] env[61970]: INFO nova.compute.manager [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] instance snapshotting [ 962.119788] env[61970]: DEBUG nova.objects.instance [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lazy-loading 'flavor' on Instance uuid ebb92a5b-f635-4504-a57f-395de514015a {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.314173] env[61970]: DEBUG nova.objects.base [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 962.317208] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40033a18-c4c9-4afb-9eb9-2a5becf01731 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.348330] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a45194e-1852-42b6-b7fc-940e99afb7ba {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.359728] env[61970]: DEBUG oslo_vmware.api [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 962.359728] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d4b946-dacb-c0a3-ac50-f9b2df2f2ad6" [ 962.359728] env[61970]: _type = "Task" [ 962.359728] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.369986] env[61970]: DEBUG oslo_vmware.api [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d4b946-dacb-c0a3-ac50-f9b2df2f2ad6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.427149] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-821a87c2-68bd-4107-9990-845302d72a92 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.434764] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a8bafb-d8cc-4d36-88d6-6885ee1644fc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.470109] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519f5435-e0d4-4fa6-a8af-17207af763ed {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.479334] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3820417e-3159-43ca-aa5e-a19b38454074 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.495153] env[61970]: DEBUG nova.compute.provider_tree [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.519777] env[61970]: DEBUG nova.compute.manager [req-bd940de8-7e54-49c4-a24e-4147d8b7e23c req-94383626-d902-4833-8eac-d2b7ce5a3264 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Received event network-changed-5b506f12-7feb-41e2-8e18-abc9262ac87b {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 962.520311] env[61970]: DEBUG nova.compute.manager [req-bd940de8-7e54-49c4-a24e-4147d8b7e23c req-94383626-d902-4833-8eac-d2b7ce5a3264 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Refreshing instance network info cache due to event network-changed-5b506f12-7feb-41e2-8e18-abc9262ac87b. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 962.521523] env[61970]: DEBUG oslo_concurrency.lockutils [req-bd940de8-7e54-49c4-a24e-4147d8b7e23c req-94383626-d902-4833-8eac-d2b7ce5a3264 service nova] Acquiring lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.522863] env[61970]: DEBUG oslo_concurrency.lockutils [req-bd940de8-7e54-49c4-a24e-4147d8b7e23c req-94383626-d902-4833-8eac-d2b7ce5a3264 service nova] Acquired lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.522863] env[61970]: DEBUG nova.network.neutron [req-bd940de8-7e54-49c4-a24e-4147d8b7e23c req-94383626-d902-4833-8eac-d2b7ce5a3264 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Refreshing network info cache for port 5b506f12-7feb-41e2-8e18-abc9262ac87b {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 962.626867] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33cd931a-3349-49b5-a84d-c73ea626a70b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.651312] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8405a3d-7cc1-442c-92af-bcc0e9a918a7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.872509] env[61970]: DEBUG oslo_vmware.api [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d4b946-dacb-c0a3-ac50-f9b2df2f2ad6, 'name': SearchDatastore_Task, 'duration_secs': 0.008413} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.872509] env[61970]: DEBUG oslo_concurrency.lockutils [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.000700] env[61970]: DEBUG nova.scheduler.client.report [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 963.165400] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Creating Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 963.165754] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a6655fee-7b16-4c0d-8655-8c1f9f810f14 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.176185] env[61970]: DEBUG oslo_vmware.api [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 963.176185] env[61970]: value = "task-1355941" [ 963.176185] env[61970]: _type = "Task" [ 963.176185] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.184359] env[61970]: DEBUG oslo_vmware.api [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355941, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.484323] env[61970]: DEBUG nova.network.neutron [req-bd940de8-7e54-49c4-a24e-4147d8b7e23c req-94383626-d902-4833-8eac-d2b7ce5a3264 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Updated VIF entry in instance network info cache for port 5b506f12-7feb-41e2-8e18-abc9262ac87b. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 963.484716] env[61970]: DEBUG nova.network.neutron [req-bd940de8-7e54-49c4-a24e-4147d8b7e23c req-94383626-d902-4833-8eac-d2b7ce5a3264 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Updating instance_info_cache with network_info: [{"id": "5b506f12-7feb-41e2-8e18-abc9262ac87b", "address": "fa:16:3e:f9:43:05", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b506f12-7f", "ovs_interfaceid": "5b506f12-7feb-41e2-8e18-abc9262ac87b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.512812] env[61970]: DEBUG oslo_concurrency.lockutils [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.332s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.512812] env[61970]: DEBUG nova.compute.manager [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 963.515532] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.703s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.517745] env[61970]: INFO nova.compute.claims [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 963.688545] env[61970]: DEBUG oslo_vmware.api [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355941, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.988057] env[61970]: DEBUG oslo_concurrency.lockutils [req-bd940de8-7e54-49c4-a24e-4147d8b7e23c req-94383626-d902-4833-8eac-d2b7ce5a3264 service nova] Releasing lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.016887] env[61970]: DEBUG nova.compute.utils [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 964.018524] env[61970]: DEBUG nova.compute.manager [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 964.019222] env[61970]: DEBUG nova.network.neutron [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 964.057098] env[61970]: DEBUG nova.policy [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd04344afddb14418b5018b69fbd7694e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c6a4997482a4834a30c79c4cbab2cc1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 964.189438] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.189438] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.190771] env[61970]: DEBUG oslo_vmware.api [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355941, 'name': CreateSnapshot_Task, 'duration_secs': 0.713255} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.190771] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Created Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 964.191555] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd51a87d-baa3-450e-9012-87cf8ec9073a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.323254] env[61970]: DEBUG nova.network.neutron [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Successfully created port: eb47f433-6083-4086-ac69-1df8051e0d2f {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 964.522505] env[61970]: DEBUG nova.compute.manager [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 964.572254] env[61970]: DEBUG nova.compute.manager [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Received event network-changed-4434695d-ddda-4840-b1da-8159c98dcae0 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.572430] env[61970]: DEBUG nova.compute.manager [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Refreshing instance network info cache due to event network-changed-4434695d-ddda-4840-b1da-8159c98dcae0. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 964.572631] env[61970]: DEBUG oslo_concurrency.lockutils [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] Acquiring lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.573202] env[61970]: DEBUG oslo_concurrency.lockutils [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] Acquired lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.573202] env[61970]: DEBUG nova.network.neutron [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Refreshing network info cache for port 4434695d-ddda-4840-b1da-8159c98dcae0 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 964.691050] env[61970]: DEBUG nova.compute.manager [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 964.710278] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Creating linked-clone VM from snapshot {{(pid=61970) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 964.713391] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3ea6bae8-2590-45e7-b5ef-0ad65b0563cd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.725023] env[61970]: DEBUG oslo_vmware.api [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 964.725023] env[61970]: value = "task-1355942" [ 964.725023] env[61970]: _type = "Task" [ 964.725023] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.733019] env[61970]: DEBUG oslo_vmware.api [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355942, 'name': CloneVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.735482] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-306f36a6-369f-42d7-8dc9-ac0666e2a342 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.743184] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-961fa9cc-4270-4013-a500-c1755db2748d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.774709] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b833bc-263c-4d79-9db8-b5f1724a2b8d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.784033] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51902e38-66f6-4057-96fd-5f24f4957ae3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.800260] env[61970]: DEBUG nova.compute.provider_tree [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.211847] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.234685] env[61970]: DEBUG oslo_vmware.api [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355942, 'name': CloneVM_Task} progress is 94%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.283617] env[61970]: DEBUG nova.network.neutron [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Updated VIF entry in instance network info cache for port 4434695d-ddda-4840-b1da-8159c98dcae0. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 965.283998] env[61970]: DEBUG nova.network.neutron [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Updating instance_info_cache with network_info: [{"id": "4434695d-ddda-4840-b1da-8159c98dcae0", "address": "fa:16:3e:85:27:97", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4434695d-dd", "ovs_interfaceid": "4434695d-ddda-4840-b1da-8159c98dcae0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.303862] env[61970]: DEBUG nova.scheduler.client.report [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 965.535674] env[61970]: DEBUG nova.compute.manager [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 965.559261] env[61970]: DEBUG nova.virt.hardware [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 965.559685] env[61970]: DEBUG nova.virt.hardware [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 965.559760] env[61970]: DEBUG nova.virt.hardware [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 965.559924] env[61970]: DEBUG nova.virt.hardware [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 965.560090] env[61970]: DEBUG nova.virt.hardware [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 965.560247] env[61970]: DEBUG nova.virt.hardware [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 965.560462] env[61970]: DEBUG nova.virt.hardware [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 965.560668] env[61970]: DEBUG nova.virt.hardware [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 965.560856] env[61970]: DEBUG nova.virt.hardware [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 965.561032] env[61970]: DEBUG nova.virt.hardware [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 965.561215] env[61970]: DEBUG nova.virt.hardware [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 965.562122] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc4fbb7-eeca-4d50-ba9a-f1ea11c82fcb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.570855] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c34e66-36ed-40be-ad03-cbfa21a978df {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.734655] env[61970]: DEBUG oslo_vmware.api [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355942, 'name': CloneVM_Task} progress is 94%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.787058] env[61970]: DEBUG oslo_concurrency.lockutils [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] Releasing lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.787438] env[61970]: DEBUG nova.compute.manager [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Received event network-changed-4434695d-ddda-4840-b1da-8159c98dcae0 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.787690] env[61970]: DEBUG nova.compute.manager [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Refreshing instance network info cache due to event network-changed-4434695d-ddda-4840-b1da-8159c98dcae0. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 965.791024] env[61970]: DEBUG oslo_concurrency.lockutils [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] Acquiring lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.791024] env[61970]: DEBUG oslo_concurrency.lockutils [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] Acquired lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.791024] env[61970]: DEBUG nova.network.neutron [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Refreshing network info cache for port 4434695d-ddda-4840-b1da-8159c98dcae0 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 965.809241] env[61970]: DEBUG nova.compute.manager [req-e100134d-938f-445e-a013-3c08d52d35c2 req-3607678d-9980-4058-96cb-71fa165e8ca2 service nova] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Received event network-vif-plugged-eb47f433-6083-4086-ac69-1df8051e0d2f {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.809688] env[61970]: DEBUG oslo_concurrency.lockutils [req-e100134d-938f-445e-a013-3c08d52d35c2 req-3607678d-9980-4058-96cb-71fa165e8ca2 service nova] Acquiring lock "479342e0-7627-4dda-a589-a7bc5a5c4270-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.809778] env[61970]: DEBUG oslo_concurrency.lockutils [req-e100134d-938f-445e-a013-3c08d52d35c2 req-3607678d-9980-4058-96cb-71fa165e8ca2 service nova] Lock "479342e0-7627-4dda-a589-a7bc5a5c4270-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.809983] env[61970]: DEBUG oslo_concurrency.lockutils [req-e100134d-938f-445e-a013-3c08d52d35c2 req-3607678d-9980-4058-96cb-71fa165e8ca2 service nova] Lock "479342e0-7627-4dda-a589-a7bc5a5c4270-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.813421] env[61970]: DEBUG nova.compute.manager [req-e100134d-938f-445e-a013-3c08d52d35c2 req-3607678d-9980-4058-96cb-71fa165e8ca2 service nova] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] No waiting events found dispatching network-vif-plugged-eb47f433-6083-4086-ac69-1df8051e0d2f {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 965.813647] env[61970]: WARNING nova.compute.manager [req-e100134d-938f-445e-a013-3c08d52d35c2 req-3607678d-9980-4058-96cb-71fa165e8ca2 service nova] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Received unexpected event network-vif-plugged-eb47f433-6083-4086-ac69-1df8051e0d2f for instance with vm_state building and task_state spawning. [ 965.814731] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.300s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.815228] env[61970]: DEBUG nova.compute.manager [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 965.818607] env[61970]: DEBUG oslo_concurrency.lockutils [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.947s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.944454] env[61970]: DEBUG nova.network.neutron [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Successfully updated port: eb47f433-6083-4086-ac69-1df8051e0d2f {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 966.234745] env[61970]: DEBUG oslo_vmware.api [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355942, 'name': CloneVM_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.320271] env[61970]: DEBUG nova.compute.utils [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 966.321782] env[61970]: DEBUG nova.compute.manager [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 966.321920] env[61970]: DEBUG nova.network.neutron [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 966.389639] env[61970]: DEBUG nova.policy [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64082edf5f2e4621b5f3586e1792e42a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9235eeeb6fa4ccf959ffbea456f3694', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 966.450945] env[61970]: DEBUG oslo_concurrency.lockutils [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "refresh_cache-479342e0-7627-4dda-a589-a7bc5a5c4270" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.450945] env[61970]: DEBUG oslo_concurrency.lockutils [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "refresh_cache-479342e0-7627-4dda-a589-a7bc5a5c4270" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.450945] env[61970]: DEBUG nova.network.neutron [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 966.511835] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25b2463-c882-44a4-a0af-556b6d668486 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.520049] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c953a90-3dd6-4aa6-99c7-d9136c3cbf4a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.550504] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dc1b7c7-c79f-4162-b875-f395209ebec0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.560376] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af99521a-22d1-47f0-84b0-f30dae750575 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.574148] env[61970]: DEBUG nova.compute.provider_tree [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.613036] env[61970]: DEBUG nova.network.neutron [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Updated VIF entry in instance network info cache for port 4434695d-ddda-4840-b1da-8159c98dcae0. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 966.613460] env[61970]: DEBUG nova.network.neutron [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Updating instance_info_cache with network_info: [{"id": "4434695d-ddda-4840-b1da-8159c98dcae0", "address": "fa:16:3e:85:27:97", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4434695d-dd", "ovs_interfaceid": "4434695d-ddda-4840-b1da-8159c98dcae0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.736015] env[61970]: DEBUG oslo_vmware.api [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355942, 'name': CloneVM_Task, 'duration_secs': 1.525972} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.736398] env[61970]: INFO nova.virt.vmwareapi.vmops [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Created linked-clone VM from snapshot [ 966.737970] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0299236e-0cdd-49e8-9ce7-0071ae34b0b2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.745772] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Uploading image 1a7f3a54-27a5-49f0-9a72-74cdc8f4f51b {{(pid=61970) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 966.768018] env[61970]: DEBUG nova.network.neutron [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Successfully created port: f73342d3-c298-4151-9bba-79202180cd9a {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 966.771876] env[61970]: DEBUG oslo_vmware.rw_handles [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 966.771876] env[61970]: value = "vm-288848" [ 966.771876] env[61970]: _type = "VirtualMachine" [ 966.771876] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 966.773910] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0541997d-c915-438e-802c-8e7a60f5f824 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.779413] env[61970]: DEBUG oslo_vmware.rw_handles [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lease: (returnval){ [ 966.779413] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52760780-d5f5-e57d-9118-ae3bdc92da08" [ 966.779413] env[61970]: _type = "HttpNfcLease" [ 966.779413] env[61970]: } obtained for exporting VM: (result){ [ 966.779413] env[61970]: value = "vm-288848" [ 966.779413] env[61970]: _type = "VirtualMachine" [ 966.779413] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 966.779777] env[61970]: DEBUG oslo_vmware.api [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the lease: (returnval){ [ 966.779777] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52760780-d5f5-e57d-9118-ae3bdc92da08" [ 966.779777] env[61970]: _type = "HttpNfcLease" [ 966.779777] env[61970]: } to be ready. {{(pid=61970) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 966.786357] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 966.786357] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52760780-d5f5-e57d-9118-ae3bdc92da08" [ 966.786357] env[61970]: _type = "HttpNfcLease" [ 966.786357] env[61970]: } is initializing. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 966.827392] env[61970]: DEBUG nova.compute.manager [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 966.996532] env[61970]: DEBUG nova.network.neutron [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 967.077639] env[61970]: DEBUG nova.scheduler.client.report [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 967.116456] env[61970]: DEBUG oslo_concurrency.lockutils [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] Releasing lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.116727] env[61970]: DEBUG nova.compute.manager [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Received event network-changed-5b506f12-7feb-41e2-8e18-abc9262ac87b {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.116901] env[61970]: DEBUG nova.compute.manager [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Refreshing instance network info cache due to event network-changed-5b506f12-7feb-41e2-8e18-abc9262ac87b. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 967.117135] env[61970]: DEBUG oslo_concurrency.lockutils [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] Acquiring lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.117280] env[61970]: DEBUG oslo_concurrency.lockutils [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] Acquired lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.117516] env[61970]: DEBUG nova.network.neutron [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Refreshing network info cache for port 5b506f12-7feb-41e2-8e18-abc9262ac87b {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 967.220217] env[61970]: DEBUG nova.network.neutron [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Updating instance_info_cache with network_info: [{"id": "eb47f433-6083-4086-ac69-1df8051e0d2f", "address": "fa:16:3e:bb:33:a4", "network": {"id": "62edd125-b77c-4418-a465-1dcb3a0733c1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-371307366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6a4997482a4834a30c79c4cbab2cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb47f433-60", "ovs_interfaceid": "eb47f433-6083-4086-ac69-1df8051e0d2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.287828] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 967.287828] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52760780-d5f5-e57d-9118-ae3bdc92da08" [ 967.287828] env[61970]: _type = "HttpNfcLease" [ 967.287828] env[61970]: } is ready. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 967.288222] env[61970]: DEBUG oslo_vmware.rw_handles [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 967.288222] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52760780-d5f5-e57d-9118-ae3bdc92da08" [ 967.288222] env[61970]: _type = "HttpNfcLease" [ 967.288222] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 967.288884] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ee902f0-6d96-46ff-8696-1b3a6d9c98db {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.296195] env[61970]: DEBUG oslo_vmware.rw_handles [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522eedc8-b148-74f5-aa9d-18181a455731/disk-0.vmdk from lease info. {{(pid=61970) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 967.296416] env[61970]: DEBUG oslo_vmware.rw_handles [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522eedc8-b148-74f5-aa9d-18181a455731/disk-0.vmdk for reading. {{(pid=61970) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 967.393156] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f3126f96-1d0e-4944-b338-b4277732e2e2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.723056] env[61970]: DEBUG oslo_concurrency.lockutils [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "refresh_cache-479342e0-7627-4dda-a589-a7bc5a5c4270" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.723358] env[61970]: DEBUG nova.compute.manager [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Instance network_info: |[{"id": "eb47f433-6083-4086-ac69-1df8051e0d2f", "address": "fa:16:3e:bb:33:a4", "network": {"id": "62edd125-b77c-4418-a465-1dcb3a0733c1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-371307366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6a4997482a4834a30c79c4cbab2cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb47f433-60", "ovs_interfaceid": "eb47f433-6083-4086-ac69-1df8051e0d2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 967.724171] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:33:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eb47f433-6083-4086-ac69-1df8051e0d2f', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 967.732017] env[61970]: DEBUG oslo.service.loopingcall [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.732628] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 967.732998] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e58188cb-516b-4d1c-96b0-5d77d22d69f6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.757046] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 967.757046] env[61970]: value = "task-1355944" [ 967.757046] env[61970]: _type = "Task" [ 967.757046] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.765911] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355944, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.851695] env[61970]: DEBUG nova.compute.manager [req-20d82ad5-9eaa-45cc-8498-b792a51655ce req-d7672d09-901e-4aff-99de-6ff068785aff service nova] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Received event network-changed-eb47f433-6083-4086-ac69-1df8051e0d2f {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.851912] env[61970]: DEBUG nova.compute.manager [req-20d82ad5-9eaa-45cc-8498-b792a51655ce req-d7672d09-901e-4aff-99de-6ff068785aff service nova] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Refreshing instance network info cache due to event network-changed-eb47f433-6083-4086-ac69-1df8051e0d2f. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 967.852637] env[61970]: DEBUG oslo_concurrency.lockutils [req-20d82ad5-9eaa-45cc-8498-b792a51655ce req-d7672d09-901e-4aff-99de-6ff068785aff service nova] Acquiring lock "refresh_cache-479342e0-7627-4dda-a589-a7bc5a5c4270" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.852944] env[61970]: DEBUG oslo_concurrency.lockutils [req-20d82ad5-9eaa-45cc-8498-b792a51655ce req-d7672d09-901e-4aff-99de-6ff068785aff service nova] Acquired lock "refresh_cache-479342e0-7627-4dda-a589-a7bc5a5c4270" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.853516] env[61970]: DEBUG nova.network.neutron [req-20d82ad5-9eaa-45cc-8498-b792a51655ce req-d7672d09-901e-4aff-99de-6ff068785aff service nova] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Refreshing network info cache for port eb47f433-6083-4086-ac69-1df8051e0d2f {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 967.859699] env[61970]: DEBUG nova.compute.manager [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 967.930167] env[61970]: DEBUG nova.network.neutron [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Updated VIF entry in instance network info cache for port 5b506f12-7feb-41e2-8e18-abc9262ac87b. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 967.930827] env[61970]: DEBUG nova.network.neutron [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Updating instance_info_cache with network_info: [{"id": "5b506f12-7feb-41e2-8e18-abc9262ac87b", "address": "fa:16:3e:f9:43:05", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b506f12-7f", "ovs_interfaceid": "5b506f12-7feb-41e2-8e18-abc9262ac87b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.088994] env[61970]: DEBUG oslo_concurrency.lockutils [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.270s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.091893] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.880s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.093803] env[61970]: INFO nova.compute.claims [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 968.280954] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355944, 'name': CreateVM_Task} progress is 25%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.433886] env[61970]: DEBUG oslo_concurrency.lockutils [req-ad9dec3a-fed2-441f-8b6e-b03a85542ef3 req-66e0904e-688a-4f49-b584-f875db1d94b0 service nova] Releasing lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.454208] env[61970]: DEBUG nova.virt.hardware [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 968.454553] env[61970]: DEBUG nova.virt.hardware [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 968.454796] env[61970]: DEBUG nova.virt.hardware [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 968.455010] env[61970]: DEBUG nova.virt.hardware [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 968.455249] env[61970]: DEBUG nova.virt.hardware [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 968.455459] env[61970]: DEBUG nova.virt.hardware [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 968.456102] env[61970]: DEBUG nova.virt.hardware [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 968.456102] env[61970]: DEBUG nova.virt.hardware [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 968.456225] env[61970]: DEBUG nova.virt.hardware [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 968.456390] env[61970]: DEBUG nova.virt.hardware [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 968.456605] env[61970]: DEBUG nova.virt.hardware [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 968.458922] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7e9784-32a9-44dd-9c9c-4e165902b950 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.468864] env[61970]: DEBUG oslo_vmware.rw_handles [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526053fe-3aff-2103-cc95-816f32571d5a/disk-0.vmdk. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 968.469992] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ffd3a2f-c942-4281-a6c7-a163c90f7727 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.475474] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e09ba8-83de-42e2-934f-58e7308b6da0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.492793] env[61970]: DEBUG oslo_vmware.rw_handles [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526053fe-3aff-2103-cc95-816f32571d5a/disk-0.vmdk is in state: ready. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 968.492983] env[61970]: ERROR oslo_vmware.rw_handles [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526053fe-3aff-2103-cc95-816f32571d5a/disk-0.vmdk due to incomplete transfer. [ 968.493299] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-dcae9201-aef9-41b7-989d-d39771eee4b8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.501474] env[61970]: DEBUG oslo_vmware.rw_handles [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526053fe-3aff-2103-cc95-816f32571d5a/disk-0.vmdk. {{(pid=61970) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 968.501707] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Uploaded image 8ff34e61-3203-40b2-805a-fd5355a36d65 to the Glance image server {{(pid=61970) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 968.504597] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Destroying the VM {{(pid=61970) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 968.504894] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c347c827-f70c-4314-bdf7-bc83925e2b8f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.511882] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 968.511882] env[61970]: value = "task-1355945" [ 968.511882] env[61970]: _type = "Task" [ 968.511882] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.524143] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355945, 'name': Destroy_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.653834] env[61970]: DEBUG nova.network.neutron [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Successfully updated port: f73342d3-c298-4151-9bba-79202180cd9a {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 968.682521] env[61970]: INFO nova.scheduler.client.report [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Deleted allocation for migration a24105dc-3e41-4a23-aea8-b524dd9b68a8 [ 968.718316] env[61970]: DEBUG nova.network.neutron [req-20d82ad5-9eaa-45cc-8498-b792a51655ce req-d7672d09-901e-4aff-99de-6ff068785aff service nova] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Updated VIF entry in instance network info cache for port eb47f433-6083-4086-ac69-1df8051e0d2f. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 968.719047] env[61970]: DEBUG nova.network.neutron [req-20d82ad5-9eaa-45cc-8498-b792a51655ce req-d7672d09-901e-4aff-99de-6ff068785aff service nova] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Updating instance_info_cache with network_info: [{"id": "eb47f433-6083-4086-ac69-1df8051e0d2f", "address": "fa:16:3e:bb:33:a4", "network": {"id": "62edd125-b77c-4418-a465-1dcb3a0733c1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-371307366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6a4997482a4834a30c79c4cbab2cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb47f433-60", "ovs_interfaceid": "eb47f433-6083-4086-ac69-1df8051e0d2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.767972] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355944, 'name': CreateVM_Task} progress is 25%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.023070] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355945, 'name': Destroy_Task, 'duration_secs': 0.322813} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.023482] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Destroyed the VM [ 969.023799] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Deleting Snapshot of the VM instance {{(pid=61970) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 969.024133] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-120b4a94-2dd3-4618-b605-3dd092fc3ebc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.030960] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 969.030960] env[61970]: value = "task-1355946" [ 969.030960] env[61970]: _type = "Task" [ 969.030960] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.038950] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355946, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.156458] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "refresh_cache-4d093692-07d3-49aa-8b0a-fb52bd9c0ffc" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.156603] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "refresh_cache-4d093692-07d3-49aa-8b0a-fb52bd9c0ffc" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.156752] env[61970]: DEBUG nova.network.neutron [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 969.166391] env[61970]: DEBUG oslo_concurrency.lockutils [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "dc38f9da-9f82-4707-9764-b28aa83a8ece" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.190064] env[61970]: DEBUG oslo_concurrency.lockutils [None req-28e0ed20-e919-42f7-aff9-55a659e44acd tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "dc38f9da-9f82-4707-9764-b28aa83a8ece" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.702s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.191530] env[61970]: DEBUG oslo_concurrency.lockutils [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "dc38f9da-9f82-4707-9764-b28aa83a8ece" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.025s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.191675] env[61970]: DEBUG oslo_concurrency.lockutils [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "dc38f9da-9f82-4707-9764-b28aa83a8ece-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.191984] env[61970]: DEBUG oslo_concurrency.lockutils [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "dc38f9da-9f82-4707-9764-b28aa83a8ece-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.192309] env[61970]: DEBUG oslo_concurrency.lockutils [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "dc38f9da-9f82-4707-9764-b28aa83a8ece-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.197904] env[61970]: INFO nova.compute.manager [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Terminating instance [ 969.203051] env[61970]: DEBUG nova.compute.manager [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 969.203330] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 969.204288] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-154b425e-1be7-4c18-aba1-b3bded8e196a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.213691] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 969.214060] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-82935d02-b34e-44c5-b05e-5b6eb4be5c6e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.220721] env[61970]: DEBUG oslo_vmware.api [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 969.220721] env[61970]: value = "task-1355947" [ 969.220721] env[61970]: _type = "Task" [ 969.220721] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.226705] env[61970]: DEBUG oslo_concurrency.lockutils [req-20d82ad5-9eaa-45cc-8498-b792a51655ce req-d7672d09-901e-4aff-99de-6ff068785aff service nova] Releasing lock "refresh_cache-479342e0-7627-4dda-a589-a7bc5a5c4270" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.234880] env[61970]: DEBUG oslo_vmware.api [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355947, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.271693] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355944, 'name': CreateVM_Task, 'duration_secs': 1.194784} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.271956] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 969.272788] env[61970]: DEBUG oslo_concurrency.lockutils [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.272959] env[61970]: DEBUG oslo_concurrency.lockutils [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.273352] env[61970]: DEBUG oslo_concurrency.lockutils [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 969.273671] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da861f48-5d57-4da4-bd09-858672ce6ef6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.279185] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 969.279185] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52431467-53f2-20be-97f0-bf9c901ebc25" [ 969.279185] env[61970]: _type = "Task" [ 969.279185] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.283998] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-623a9d89-07bc-4afe-abfc-23229b907662 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.292058] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52431467-53f2-20be-97f0-bf9c901ebc25, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.294827] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0073fe45-80d5-4cf7-a661-af6a8c4bd9e5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.325896] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8636cbf-51b2-4994-9e4f-075c53d2f185 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.335391] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f60037c2-f76b-47c1-b5cb-f6c0f98e5c73 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.350038] env[61970]: DEBUG nova.compute.provider_tree [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.540792] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355946, 'name': RemoveSnapshot_Task, 'duration_secs': 0.388379} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.541205] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Deleted Snapshot of the VM instance {{(pid=61970) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 969.541393] env[61970]: DEBUG nova.compute.manager [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 969.542164] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cef85d63-89fa-44f0-89bf-56cbfd5f8861 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.665900] env[61970]: DEBUG oslo_concurrency.lockutils [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "3b1e8446-7862-4532-b2ed-611b63dbf391" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.666178] env[61970]: DEBUG oslo_concurrency.lockutils [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "3b1e8446-7862-4532-b2ed-611b63dbf391" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.697112] env[61970]: DEBUG nova.network.neutron [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 969.731517] env[61970]: DEBUG oslo_vmware.api [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355947, 'name': PowerOffVM_Task, 'duration_secs': 0.196667} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.731856] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 969.732082] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 969.732449] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-989e9bb3-1e43-414c-916b-9e0336304c73 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.789715] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52431467-53f2-20be-97f0-bf9c901ebc25, 'name': SearchDatastore_Task, 'duration_secs': 0.012775} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.790207] env[61970]: DEBUG oslo_concurrency.lockutils [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.790339] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 969.790576] env[61970]: DEBUG oslo_concurrency.lockutils [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.790777] env[61970]: DEBUG oslo_concurrency.lockutils [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.790981] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 969.791285] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-57887bfb-8214-48ad-ab96-93738c449270 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.798648] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 969.799346] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 969.799346] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Deleting the datastore file [datastore1] dc38f9da-9f82-4707-9764-b28aa83a8ece {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 969.800587] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-66f6532c-2dff-4967-89b7-63b49ec2b371 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.802486] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 969.802667] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 969.803405] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92eaf911-5fbf-48b9-aca9-7be3da716641 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.808870] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 969.808870] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52779869-4d57-4319-48ef-154624995727" [ 969.808870] env[61970]: _type = "Task" [ 969.808870] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.813132] env[61970]: DEBUG oslo_vmware.api [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 969.813132] env[61970]: value = "task-1355949" [ 969.813132] env[61970]: _type = "Task" [ 969.813132] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.819340] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52779869-4d57-4319-48ef-154624995727, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.825797] env[61970]: DEBUG oslo_vmware.api [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355949, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.853043] env[61970]: DEBUG nova.scheduler.client.report [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.858058] env[61970]: DEBUG nova.network.neutron [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Updating instance_info_cache with network_info: [{"id": "f73342d3-c298-4151-9bba-79202180cd9a", "address": "fa:16:3e:68:5d:fb", "network": {"id": "e02bfa6c-8f0d-4413-8dcf-e644c07792d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1410864732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9235eeeb6fa4ccf959ffbea456f3694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf73342d3-c2", "ovs_interfaceid": "f73342d3-c298-4151-9bba-79202180cd9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.878392] env[61970]: DEBUG nova.compute.manager [req-62a7d858-d4ac-4274-bd18-0816e504e19c req-eeae6c99-9942-404b-8438-04eb330105de service nova] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Received event network-vif-plugged-f73342d3-c298-4151-9bba-79202180cd9a {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.878674] env[61970]: DEBUG oslo_concurrency.lockutils [req-62a7d858-d4ac-4274-bd18-0816e504e19c req-eeae6c99-9942-404b-8438-04eb330105de service nova] Acquiring lock "4d093692-07d3-49aa-8b0a-fb52bd9c0ffc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.878826] env[61970]: DEBUG oslo_concurrency.lockutils [req-62a7d858-d4ac-4274-bd18-0816e504e19c req-eeae6c99-9942-404b-8438-04eb330105de service nova] Lock "4d093692-07d3-49aa-8b0a-fb52bd9c0ffc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.879081] env[61970]: DEBUG oslo_concurrency.lockutils [req-62a7d858-d4ac-4274-bd18-0816e504e19c req-eeae6c99-9942-404b-8438-04eb330105de service nova] Lock "4d093692-07d3-49aa-8b0a-fb52bd9c0ffc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.879480] env[61970]: DEBUG nova.compute.manager [req-62a7d858-d4ac-4274-bd18-0816e504e19c req-eeae6c99-9942-404b-8438-04eb330105de service nova] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] No waiting events found dispatching network-vif-plugged-f73342d3-c298-4151-9bba-79202180cd9a {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 969.879480] env[61970]: WARNING nova.compute.manager [req-62a7d858-d4ac-4274-bd18-0816e504e19c req-eeae6c99-9942-404b-8438-04eb330105de service nova] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Received unexpected event network-vif-plugged-f73342d3-c298-4151-9bba-79202180cd9a for instance with vm_state building and task_state spawning. [ 969.879553] env[61970]: DEBUG nova.compute.manager [req-62a7d858-d4ac-4274-bd18-0816e504e19c req-eeae6c99-9942-404b-8438-04eb330105de service nova] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Received event network-changed-f73342d3-c298-4151-9bba-79202180cd9a {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.879719] env[61970]: DEBUG nova.compute.manager [req-62a7d858-d4ac-4274-bd18-0816e504e19c req-eeae6c99-9942-404b-8438-04eb330105de service nova] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Refreshing instance network info cache due to event network-changed-f73342d3-c298-4151-9bba-79202180cd9a. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 969.879897] env[61970]: DEBUG oslo_concurrency.lockutils [req-62a7d858-d4ac-4274-bd18-0816e504e19c req-eeae6c99-9942-404b-8438-04eb330105de service nova] Acquiring lock "refresh_cache-4d093692-07d3-49aa-8b0a-fb52bd9c0ffc" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.054445] env[61970]: INFO nova.compute.manager [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Shelve offloading [ 970.056210] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 970.056478] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e2d9fdc-a2ae-45e0-bf9a-ad11ce63cd87 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.065013] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 970.065013] env[61970]: value = "task-1355950" [ 970.065013] env[61970]: _type = "Task" [ 970.065013] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.075061] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] VM already powered off {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 970.075061] env[61970]: DEBUG nova.compute.manager [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.075452] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f21c1a-4a36-4f6d-bc30-aae5877bf461 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.081191] env[61970]: DEBUG oslo_concurrency.lockutils [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.081369] env[61970]: DEBUG oslo_concurrency.lockutils [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquired lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.081554] env[61970]: DEBUG nova.network.neutron [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 970.169180] env[61970]: DEBUG nova.compute.manager [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 970.322685] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52779869-4d57-4319-48ef-154624995727, 'name': SearchDatastore_Task, 'duration_secs': 0.010537} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.326443] env[61970]: DEBUG oslo_vmware.api [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355949, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136613} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.326665] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1230647-b06e-40bb-b9c7-b84672002c6c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.329155] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 970.329354] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 970.329537] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 970.329783] env[61970]: INFO nova.compute.manager [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Took 1.13 seconds to destroy the instance on the hypervisor. [ 970.330048] env[61970]: DEBUG oslo.service.loopingcall [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.330248] env[61970]: DEBUG nova.compute.manager [-] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 970.330349] env[61970]: DEBUG nova.network.neutron [-] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 970.334503] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 970.334503] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52dad963-ebe7-cc7d-92eb-1673abdf00e9" [ 970.334503] env[61970]: _type = "Task" [ 970.334503] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.342136] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52dad963-ebe7-cc7d-92eb-1673abdf00e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.360080] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.268s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.360662] env[61970]: DEBUG nova.compute.manager [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 970.363897] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "refresh_cache-4d093692-07d3-49aa-8b0a-fb52bd9c0ffc" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.364353] env[61970]: DEBUG nova.compute.manager [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Instance network_info: |[{"id": "f73342d3-c298-4151-9bba-79202180cd9a", "address": "fa:16:3e:68:5d:fb", "network": {"id": "e02bfa6c-8f0d-4413-8dcf-e644c07792d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1410864732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9235eeeb6fa4ccf959ffbea456f3694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf73342d3-c2", "ovs_interfaceid": "f73342d3-c298-4151-9bba-79202180cd9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 970.366783] env[61970]: DEBUG oslo_concurrency.lockutils [req-62a7d858-d4ac-4274-bd18-0816e504e19c req-eeae6c99-9942-404b-8438-04eb330105de service nova] Acquired lock "refresh_cache-4d093692-07d3-49aa-8b0a-fb52bd9c0ffc" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.366974] env[61970]: DEBUG nova.network.neutron [req-62a7d858-d4ac-4274-bd18-0816e504e19c req-eeae6c99-9942-404b-8438-04eb330105de service nova] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Refreshing network info cache for port f73342d3-c298-4151-9bba-79202180cd9a {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 970.368606] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:5d:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '510d3c47-3615-43d5-aa5d-a279fd915e71', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f73342d3-c298-4151-9bba-79202180cd9a', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 970.376745] env[61970]: DEBUG oslo.service.loopingcall [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.377773] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 970.378348] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2a61defa-5d4a-4051-982f-7b0a5e97385e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.400967] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 970.400967] env[61970]: value = "task-1355951" [ 970.400967] env[61970]: _type = "Task" [ 970.400967] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.409088] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355951, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.700143] env[61970]: DEBUG oslo_concurrency.lockutils [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.700570] env[61970]: DEBUG oslo_concurrency.lockutils [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.702298] env[61970]: INFO nova.compute.claims [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 970.807352] env[61970]: DEBUG nova.network.neutron [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Updating instance_info_cache with network_info: [{"id": "5c475b00-f4ed-4e09-916c-6f3ca7845784", "address": "fa:16:3e:27:f3:40", "network": {"id": "70081995-3d6c-466e-b811-1c8e532e552f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-641102750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1d4724cbb0841859369af01f2c03db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aaf1b231-3660-4453-b4f3-44d825b9a5dd", "external-id": "nsx-vlan-transportzone-6", "segmentation_id": 6, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c475b00-f4", "ovs_interfaceid": "5c475b00-f4ed-4e09-916c-6f3ca7845784", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.844890] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52dad963-ebe7-cc7d-92eb-1673abdf00e9, 'name': SearchDatastore_Task, 'duration_secs': 0.008639} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.845185] env[61970]: DEBUG oslo_concurrency.lockutils [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.845461] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] 479342e0-7627-4dda-a589-a7bc5a5c4270/479342e0-7627-4dda-a589-a7bc5a5c4270.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 970.845764] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-13ed487b-1773-4eab-8938-2d95d96e792c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.852658] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 970.852658] env[61970]: value = "task-1355952" [ 970.852658] env[61970]: _type = "Task" [ 970.852658] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.861423] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355952, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.869035] env[61970]: DEBUG nova.compute.utils [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 970.874385] env[61970]: DEBUG nova.compute.manager [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 970.874385] env[61970]: DEBUG nova.network.neutron [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 970.917022] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355951, 'name': CreateVM_Task, 'duration_secs': 0.446748} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.917022] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 970.917022] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.917022] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.917022] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 970.917022] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69706d82-b7f7-4d6e-bfe8-2032763a04a1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.921703] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 970.921703] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5258469e-6dab-fb1b-54d5-d2b86f61bee7" [ 970.921703] env[61970]: _type = "Task" [ 970.921703] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.931787] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5258469e-6dab-fb1b-54d5-d2b86f61bee7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.951184] env[61970]: DEBUG nova.policy [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '223ba28a2a204f8ba9e6d2e0f876ddd2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '28a078b8af9c469eb279be4da7459166', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 971.070984] env[61970]: DEBUG nova.network.neutron [-] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.310557] env[61970]: DEBUG oslo_concurrency.lockutils [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Releasing lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.365509] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355952, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490511} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.365509] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] 479342e0-7627-4dda-a589-a7bc5a5c4270/479342e0-7627-4dda-a589-a7bc5a5c4270.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 971.365509] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 971.365509] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-491abd52-e1b1-47fa-9161-fce46db16c48 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.370520] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 971.370520] env[61970]: value = "task-1355953" [ 971.370520] env[61970]: _type = "Task" [ 971.370520] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.378947] env[61970]: DEBUG nova.compute.manager [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 971.381960] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355953, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.433481] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5258469e-6dab-fb1b-54d5-d2b86f61bee7, 'name': SearchDatastore_Task, 'duration_secs': 0.012062} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.433773] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.433997] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 971.434244] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.434386] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.434555] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 971.434824] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-518d2b42-11e9-44a8-bf9a-57b2749a4446 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.443671] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 971.443865] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 971.447801] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b445fac-47ed-41a6-8010-3cbb771f0b4e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.449644] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 971.449644] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52013e78-2876-91f1-52ac-b5993c91c7a0" [ 971.449644] env[61970]: _type = "Task" [ 971.449644] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.458116] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52013e78-2876-91f1-52ac-b5993c91c7a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.513912] env[61970]: DEBUG nova.network.neutron [req-62a7d858-d4ac-4274-bd18-0816e504e19c req-eeae6c99-9942-404b-8438-04eb330105de service nova] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Updated VIF entry in instance network info cache for port f73342d3-c298-4151-9bba-79202180cd9a. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 971.513912] env[61970]: DEBUG nova.network.neutron [req-62a7d858-d4ac-4274-bd18-0816e504e19c req-eeae6c99-9942-404b-8438-04eb330105de service nova] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Updating instance_info_cache with network_info: [{"id": "f73342d3-c298-4151-9bba-79202180cd9a", "address": "fa:16:3e:68:5d:fb", "network": {"id": "e02bfa6c-8f0d-4413-8dcf-e644c07792d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1410864732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9235eeeb6fa4ccf959ffbea456f3694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf73342d3-c2", "ovs_interfaceid": "f73342d3-c298-4151-9bba-79202180cd9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.574151] env[61970]: INFO nova.compute.manager [-] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Took 1.24 seconds to deallocate network for instance. [ 971.591273] env[61970]: DEBUG nova.compute.manager [req-f12d75b9-34b6-4ec3-b723-e6e8bdf7d55f req-6aaf17c2-b59e-4681-8204-e9738cb7ae63 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Received event network-vif-unplugged-5c475b00-f4ed-4e09-916c-6f3ca7845784 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 971.591544] env[61970]: DEBUG oslo_concurrency.lockutils [req-f12d75b9-34b6-4ec3-b723-e6e8bdf7d55f req-6aaf17c2-b59e-4681-8204-e9738cb7ae63 service nova] Acquiring lock "7f59000a-94d9-45b6-aa7a-300d95793615-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.592477] env[61970]: DEBUG oslo_concurrency.lockutils [req-f12d75b9-34b6-4ec3-b723-e6e8bdf7d55f req-6aaf17c2-b59e-4681-8204-e9738cb7ae63 service nova] Lock "7f59000a-94d9-45b6-aa7a-300d95793615-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.592673] env[61970]: DEBUG oslo_concurrency.lockutils [req-f12d75b9-34b6-4ec3-b723-e6e8bdf7d55f req-6aaf17c2-b59e-4681-8204-e9738cb7ae63 service nova] Lock "7f59000a-94d9-45b6-aa7a-300d95793615-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.592855] env[61970]: DEBUG nova.compute.manager [req-f12d75b9-34b6-4ec3-b723-e6e8bdf7d55f req-6aaf17c2-b59e-4681-8204-e9738cb7ae63 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] No waiting events found dispatching network-vif-unplugged-5c475b00-f4ed-4e09-916c-6f3ca7845784 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 971.593044] env[61970]: WARNING nova.compute.manager [req-f12d75b9-34b6-4ec3-b723-e6e8bdf7d55f req-6aaf17c2-b59e-4681-8204-e9738cb7ae63 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Received unexpected event network-vif-unplugged-5c475b00-f4ed-4e09-916c-6f3ca7845784 for instance with vm_state shelved and task_state shelving_offloading. [ 971.664909] env[61970]: DEBUG nova.network.neutron [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Successfully created port: 4e65fd7e-c508-4d66-a854-d26fd273014a {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 971.719039] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 971.720055] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b27fea4-2bf2-4a53-9e97-73907cf34094 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.735090] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 971.735399] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-166e52fb-cdbd-42ff-9657-9ed371dfdcbc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.808172] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 971.808393] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 971.808576] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Deleting the datastore file [datastore2] 7f59000a-94d9-45b6-aa7a-300d95793615 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 971.808840] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bd9e8fd8-c69b-4bca-9bfd-3bfc39d1150f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.814940] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 971.814940] env[61970]: value = "task-1355955" [ 971.814940] env[61970]: _type = "Task" [ 971.814940] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.824682] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355955, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.889431] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355953, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064105} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.892415] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 971.894944] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f46832-92ba-4dfa-8951-8d941a05da59 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.922468] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 479342e0-7627-4dda-a589-a7bc5a5c4270/479342e0-7627-4dda-a589-a7bc5a5c4270.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 971.926802] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-147eecb2-ecd7-45f4-a35a-aab8a691f2f0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.942661] env[61970]: DEBUG nova.compute.manager [req-1bd9d9a8-faa3-41b0-ba9a-95ce7401c72c req-f55aa77b-2329-49d3-b3f7-32183f0a1f3a service nova] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Received event network-vif-deleted-313663a2-7d1f-40ea-abfc-c540d35235ca {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 971.952452] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 971.952452] env[61970]: value = "task-1355956" [ 971.952452] env[61970]: _type = "Task" [ 971.952452] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.959927] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b2fb66-00fc-48f7-a9d8-3f9e4354ac00 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.969321] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52013e78-2876-91f1-52ac-b5993c91c7a0, 'name': SearchDatastore_Task, 'duration_secs': 0.010915} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.969574] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355956, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.970853] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d5d0343-8572-4a54-89a1-281cecfa0e00 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.977056] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a38fa970-d51f-4714-a660-8dd3cbb43aeb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.981788] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 971.981788] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5294e02b-fb9f-3ed6-14d8-0f632103ab1d" [ 971.981788] env[61970]: _type = "Task" [ 971.981788] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.013463] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c8c5932-d421-4bde-9b1b-d3a074dd4002 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.024529] env[61970]: DEBUG oslo_concurrency.lockutils [req-62a7d858-d4ac-4274-bd18-0816e504e19c req-eeae6c99-9942-404b-8438-04eb330105de service nova] Releasing lock "refresh_cache-4d093692-07d3-49aa-8b0a-fb52bd9c0ffc" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.024529] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5294e02b-fb9f-3ed6-14d8-0f632103ab1d, 'name': SearchDatastore_Task, 'duration_secs': 0.014776} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.024529] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.024529] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc/4d093692-07d3-49aa-8b0a-fb52bd9c0ffc.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 972.024529] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3bcea8a4-42ef-4391-8996-df30d66f5502 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.030858] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e8fae25-b7d2-47b0-8651-264e77b53134 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.035976] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 972.035976] env[61970]: value = "task-1355957" [ 972.035976] env[61970]: _type = "Task" [ 972.035976] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.048715] env[61970]: DEBUG nova.compute.provider_tree [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.054968] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355957, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.082549] env[61970]: DEBUG oslo_concurrency.lockutils [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.327745] env[61970]: DEBUG oslo_vmware.api [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1355955, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176024} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.328115] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.328395] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 972.328656] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 972.349436] env[61970]: INFO nova.scheduler.client.report [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Deleted allocations for instance 7f59000a-94d9-45b6-aa7a-300d95793615 [ 972.392200] env[61970]: DEBUG nova.compute.manager [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 972.420818] env[61970]: DEBUG nova.virt.hardware [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 972.421218] env[61970]: DEBUG nova.virt.hardware [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 972.421435] env[61970]: DEBUG nova.virt.hardware [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 972.421735] env[61970]: DEBUG nova.virt.hardware [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 972.421973] env[61970]: DEBUG nova.virt.hardware [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 972.422234] env[61970]: DEBUG nova.virt.hardware [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 972.422527] env[61970]: DEBUG nova.virt.hardware [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 972.422792] env[61970]: DEBUG nova.virt.hardware [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 972.423031] env[61970]: DEBUG nova.virt.hardware [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 972.423220] env[61970]: DEBUG nova.virt.hardware [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 972.423455] env[61970]: DEBUG nova.virt.hardware [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 972.424546] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ea3c34-dada-411d-8a5d-b26da20f3068 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.434166] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76cb967e-8d3a-4f0d-87c1-cc1351491b48 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.460183] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355956, 'name': ReconfigVM_Task, 'duration_secs': 0.285464} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.460925] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 479342e0-7627-4dda-a589-a7bc5a5c4270/479342e0-7627-4dda-a589-a7bc5a5c4270.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 972.461788] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-99483f77-0442-46de-b62c-967d78e4f2c5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.470607] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 972.470607] env[61970]: value = "task-1355958" [ 972.470607] env[61970]: _type = "Task" [ 972.470607] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.482364] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355958, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.550724] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355957, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.552038] env[61970]: DEBUG nova.scheduler.client.report [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 972.685096] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 972.685363] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 972.855257] env[61970]: DEBUG oslo_concurrency.lockutils [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.981764] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355958, 'name': Rename_Task, 'duration_secs': 0.190459} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.982070] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 972.982330] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee273219-561e-4d48-9a3a-37b8f5561be7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.988657] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 972.988657] env[61970]: value = "task-1355959" [ 972.988657] env[61970]: _type = "Task" [ 972.988657] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.997101] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355959, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.053713] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355957, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.547497} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.055173] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc/4d093692-07d3-49aa-8b0a-fb52bd9c0ffc.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 973.056390] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 973.057022] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-860624f2-0365-44c8-a01a-b0107188a4ca {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.060647] env[61970]: DEBUG oslo_concurrency.lockutils [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.360s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.061362] env[61970]: DEBUG nova.compute.manager [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 973.064559] env[61970]: DEBUG oslo_concurrency.lockutils [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.982s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.065031] env[61970]: DEBUG oslo_concurrency.lockutils [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.067628] env[61970]: DEBUG oslo_concurrency.lockutils [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.213s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.067990] env[61970]: DEBUG nova.objects.instance [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lazy-loading 'resources' on Instance uuid 7f59000a-94d9-45b6-aa7a-300d95793615 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.077045] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 973.077045] env[61970]: value = "task-1355960" [ 973.077045] env[61970]: _type = "Task" [ 973.077045] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.084903] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355960, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.096017] env[61970]: INFO nova.scheduler.client.report [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Deleted allocations for instance dc38f9da-9f82-4707-9764-b28aa83a8ece [ 973.193772] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.193772] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Starting heal instance info cache {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 973.193772] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Rebuilding the list of instances to heal {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 973.256207] env[61970]: DEBUG nova.network.neutron [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Successfully updated port: 4e65fd7e-c508-4d66-a854-d26fd273014a {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 973.501113] env[61970]: DEBUG oslo_vmware.api [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355959, 'name': PowerOnVM_Task, 'duration_secs': 0.447513} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.501113] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 973.501113] env[61970]: INFO nova.compute.manager [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Took 7.97 seconds to spawn the instance on the hypervisor. [ 973.501113] env[61970]: DEBUG nova.compute.manager [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.501855] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-890031f3-0c47-4a02-86b6-50e1a51cba92 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.571715] env[61970]: DEBUG nova.objects.instance [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lazy-loading 'numa_topology' on Instance uuid 7f59000a-94d9-45b6-aa7a-300d95793615 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.574492] env[61970]: DEBUG nova.compute.utils [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 973.575477] env[61970]: DEBUG nova.compute.manager [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 973.575655] env[61970]: DEBUG nova.network.neutron [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 973.590044] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355960, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.119068} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.590044] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 973.590624] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe6765a-e3be-4f8b-a148-def32766ca60 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.617672] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc/4d093692-07d3-49aa-8b0a-fb52bd9c0ffc.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 973.618176] env[61970]: DEBUG oslo_concurrency.lockutils [None req-04b8d5b2-bd0e-47f3-89af-8d2de3485d56 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "dc38f9da-9f82-4707-9764-b28aa83a8ece" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.427s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.620220] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b58e4c90-5011-4e8a-af29-747a6ed62a06 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.637198] env[61970]: DEBUG nova.compute.manager [req-4f75be7d-fd65-4543-abdc-405d028dd3a5 req-ea1979c9-ed73-466a-baaf-c6ec4d00d3d2 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Received event network-changed-5c475b00-f4ed-4e09-916c-6f3ca7845784 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 973.637410] env[61970]: DEBUG nova.compute.manager [req-4f75be7d-fd65-4543-abdc-405d028dd3a5 req-ea1979c9-ed73-466a-baaf-c6ec4d00d3d2 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Refreshing instance network info cache due to event network-changed-5c475b00-f4ed-4e09-916c-6f3ca7845784. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 973.637686] env[61970]: DEBUG oslo_concurrency.lockutils [req-4f75be7d-fd65-4543-abdc-405d028dd3a5 req-ea1979c9-ed73-466a-baaf-c6ec4d00d3d2 service nova] Acquiring lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.637889] env[61970]: DEBUG oslo_concurrency.lockutils [req-4f75be7d-fd65-4543-abdc-405d028dd3a5 req-ea1979c9-ed73-466a-baaf-c6ec4d00d3d2 service nova] Acquired lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.638100] env[61970]: DEBUG nova.network.neutron [req-4f75be7d-fd65-4543-abdc-405d028dd3a5 req-ea1979c9-ed73-466a-baaf-c6ec4d00d3d2 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Refreshing network info cache for port 5c475b00-f4ed-4e09-916c-6f3ca7845784 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 973.646205] env[61970]: DEBUG nova.policy [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b45dc9d71c0e4693ba42322efde6c572', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5c64cb9232fb413cbd7627dcf077e9ef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 973.650026] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 973.650026] env[61970]: value = "task-1355961" [ 973.650026] env[61970]: _type = "Task" [ 973.650026] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.660132] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355961, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.698754] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 973.698959] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 973.699498] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 973.699671] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Skipping network cache update for instance because it is Building. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 973.742868] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "refresh_cache-15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.743054] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquired lock "refresh_cache-15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.743216] env[61970]: DEBUG nova.network.neutron [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Forcefully refreshing network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 973.743369] env[61970]: DEBUG nova.objects.instance [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lazy-loading 'info_cache' on Instance uuid 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.759096] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "refresh_cache-bc948a41-1a22-46fa-a76a-50b4c4afa4b7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.759559] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "refresh_cache-bc948a41-1a22-46fa-a76a-50b4c4afa4b7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.759746] env[61970]: DEBUG nova.network.neutron [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 973.942420] env[61970]: DEBUG nova.compute.manager [req-67de96e4-6464-44a9-9873-cfdb4b4cf213 req-3afb3cd8-7cfd-4129-8a16-328d5b1cb894 service nova] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Received event network-vif-plugged-4e65fd7e-c508-4d66-a854-d26fd273014a {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 973.942765] env[61970]: DEBUG oslo_concurrency.lockutils [req-67de96e4-6464-44a9-9873-cfdb4b4cf213 req-3afb3cd8-7cfd-4129-8a16-328d5b1cb894 service nova] Acquiring lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.942945] env[61970]: DEBUG oslo_concurrency.lockutils [req-67de96e4-6464-44a9-9873-cfdb4b4cf213 req-3afb3cd8-7cfd-4129-8a16-328d5b1cb894 service nova] Lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.943097] env[61970]: DEBUG oslo_concurrency.lockutils [req-67de96e4-6464-44a9-9873-cfdb4b4cf213 req-3afb3cd8-7cfd-4129-8a16-328d5b1cb894 service nova] Lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.943240] env[61970]: DEBUG nova.compute.manager [req-67de96e4-6464-44a9-9873-cfdb4b4cf213 req-3afb3cd8-7cfd-4129-8a16-328d5b1cb894 service nova] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] No waiting events found dispatching network-vif-plugged-4e65fd7e-c508-4d66-a854-d26fd273014a {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 973.943490] env[61970]: WARNING nova.compute.manager [req-67de96e4-6464-44a9-9873-cfdb4b4cf213 req-3afb3cd8-7cfd-4129-8a16-328d5b1cb894 service nova] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Received unexpected event network-vif-plugged-4e65fd7e-c508-4d66-a854-d26fd273014a for instance with vm_state building and task_state spawning. [ 973.943594] env[61970]: DEBUG nova.compute.manager [req-67de96e4-6464-44a9-9873-cfdb4b4cf213 req-3afb3cd8-7cfd-4129-8a16-328d5b1cb894 service nova] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Received event network-changed-4e65fd7e-c508-4d66-a854-d26fd273014a {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 973.943731] env[61970]: DEBUG nova.compute.manager [req-67de96e4-6464-44a9-9873-cfdb4b4cf213 req-3afb3cd8-7cfd-4129-8a16-328d5b1cb894 service nova] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Refreshing instance network info cache due to event network-changed-4e65fd7e-c508-4d66-a854-d26fd273014a. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 973.943905] env[61970]: DEBUG oslo_concurrency.lockutils [req-67de96e4-6464-44a9-9873-cfdb4b4cf213 req-3afb3cd8-7cfd-4129-8a16-328d5b1cb894 service nova] Acquiring lock "refresh_cache-bc948a41-1a22-46fa-a76a-50b4c4afa4b7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.979467] env[61970]: DEBUG nova.network.neutron [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Successfully created port: a10357cb-17f1-4e77-ae7f-ac439a804c9c {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 974.025675] env[61970]: INFO nova.compute.manager [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Took 14.44 seconds to build instance. [ 974.050765] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "7f59000a-94d9-45b6-aa7a-300d95793615" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.076745] env[61970]: DEBUG nova.compute.manager [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 974.079030] env[61970]: DEBUG nova.objects.base [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Object Instance<7f59000a-94d9-45b6-aa7a-300d95793615> lazy-loaded attributes: resources,numa_topology {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 974.160342] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355961, 'name': ReconfigVM_Task, 'duration_secs': 0.370726} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.160661] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc/4d093692-07d3-49aa-8b0a-fb52bd9c0ffc.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 974.163696] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa4b29a3-d89a-41d3-8e7a-1dab933acd5b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.170590] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 974.170590] env[61970]: value = "task-1355962" [ 974.170590] env[61970]: _type = "Task" [ 974.170590] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.183407] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355962, 'name': Rename_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.283476] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2307d7cb-09d0-4f1d-97cd-8cf9305cd385 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.295663] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36e98c48-eea3-4cf8-b0e7-fccd31e0fa0a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.342271] env[61970]: DEBUG nova.network.neutron [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 974.345112] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81919fb4-b897-4ec3-8c7c-2941e9e18995 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.354243] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-078ab249-bc8b-424e-902f-84ae83dea2f5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.368851] env[61970]: DEBUG nova.compute.provider_tree [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.429340] env[61970]: DEBUG nova.network.neutron [req-4f75be7d-fd65-4543-abdc-405d028dd3a5 req-ea1979c9-ed73-466a-baaf-c6ec4d00d3d2 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Updated VIF entry in instance network info cache for port 5c475b00-f4ed-4e09-916c-6f3ca7845784. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 974.429774] env[61970]: DEBUG nova.network.neutron [req-4f75be7d-fd65-4543-abdc-405d028dd3a5 req-ea1979c9-ed73-466a-baaf-c6ec4d00d3d2 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Updating instance_info_cache with network_info: [{"id": "5c475b00-f4ed-4e09-916c-6f3ca7845784", "address": "fa:16:3e:27:f3:40", "network": {"id": "70081995-3d6c-466e-b811-1c8e532e552f", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-641102750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1d4724cbb0841859369af01f2c03db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap5c475b00-f4", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.528490] env[61970]: DEBUG oslo_concurrency.lockutils [None req-da92c04e-78e9-4ba6-a603-f03a4192b16f tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "479342e0-7627-4dda-a589-a7bc5a5c4270" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.950s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.571733] env[61970]: DEBUG nova.network.neutron [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Updating instance_info_cache with network_info: [{"id": "4e65fd7e-c508-4d66-a854-d26fd273014a", "address": "fa:16:3e:93:cf:a0", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e65fd7e-c5", "ovs_interfaceid": "4e65fd7e-c508-4d66-a854-d26fd273014a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.680609] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355962, 'name': Rename_Task, 'duration_secs': 0.152598} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.680900] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 974.681165] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-43066d4c-c369-4839-a87e-b5ba3358ace8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.687405] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 974.687405] env[61970]: value = "task-1355963" [ 974.687405] env[61970]: _type = "Task" [ 974.687405] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.700093] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355963, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.875503] env[61970]: DEBUG nova.scheduler.client.report [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.913448] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c755e37b-22c7-4f28-8842-f81b0a8353b7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.922226] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2fbafc6e-bf49-4621-8213-5347c9c3a63d tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Suspending the VM {{(pid=61970) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 974.922584] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-70f66d19-3297-4602-9784-3c30780b3e2d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.929235] env[61970]: DEBUG oslo_vmware.api [None req-2fbafc6e-bf49-4621-8213-5347c9c3a63d tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 974.929235] env[61970]: value = "task-1355964" [ 974.929235] env[61970]: _type = "Task" [ 974.929235] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.933023] env[61970]: DEBUG oslo_concurrency.lockutils [req-4f75be7d-fd65-4543-abdc-405d028dd3a5 req-ea1979c9-ed73-466a-baaf-c6ec4d00d3d2 service nova] Releasing lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.938011] env[61970]: DEBUG oslo_vmware.api [None req-2fbafc6e-bf49-4621-8213-5347c9c3a63d tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355964, 'name': SuspendVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.074904] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "refresh_cache-bc948a41-1a22-46fa-a76a-50b4c4afa4b7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.075324] env[61970]: DEBUG nova.compute.manager [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Instance network_info: |[{"id": "4e65fd7e-c508-4d66-a854-d26fd273014a", "address": "fa:16:3e:93:cf:a0", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e65fd7e-c5", "ovs_interfaceid": "4e65fd7e-c508-4d66-a854-d26fd273014a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 975.075546] env[61970]: DEBUG oslo_concurrency.lockutils [req-67de96e4-6464-44a9-9873-cfdb4b4cf213 req-3afb3cd8-7cfd-4129-8a16-328d5b1cb894 service nova] Acquired lock "refresh_cache-bc948a41-1a22-46fa-a76a-50b4c4afa4b7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.075736] env[61970]: DEBUG nova.network.neutron [req-67de96e4-6464-44a9-9873-cfdb4b4cf213 req-3afb3cd8-7cfd-4129-8a16-328d5b1cb894 service nova] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Refreshing network info cache for port 4e65fd7e-c508-4d66-a854-d26fd273014a {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 975.077250] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:cf:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b356db78-99c7-4464-822c-fc7e193f7878', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4e65fd7e-c508-4d66-a854-d26fd273014a', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 975.088708] env[61970]: DEBUG oslo.service.loopingcall [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 975.089979] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 975.091248] env[61970]: DEBUG nova.compute.manager [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 975.093718] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-52183cca-3d55-4bbc-b96d-6ab073fc2eea {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.119050] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 975.119050] env[61970]: value = "task-1355965" [ 975.119050] env[61970]: _type = "Task" [ 975.119050] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.121480] env[61970]: DEBUG nova.virt.hardware [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 975.121736] env[61970]: DEBUG nova.virt.hardware [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 975.121901] env[61970]: DEBUG nova.virt.hardware [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 975.122106] env[61970]: DEBUG nova.virt.hardware [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 975.122278] env[61970]: DEBUG nova.virt.hardware [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 975.122424] env[61970]: DEBUG nova.virt.hardware [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 975.122636] env[61970]: DEBUG nova.virt.hardware [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 975.122802] env[61970]: DEBUG nova.virt.hardware [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 975.122970] env[61970]: DEBUG nova.virt.hardware [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 975.123156] env[61970]: DEBUG nova.virt.hardware [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 975.123333] env[61970]: DEBUG nova.virt.hardware [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 975.124280] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4080ff-f961-4556-aa4b-7c1fdd7cc125 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.136018] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355965, 'name': CreateVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.139347] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb0a155-f246-41f7-b00e-87656dcc0705 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.200601] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355963, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.225541] env[61970]: DEBUG oslo_vmware.rw_handles [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522eedc8-b148-74f5-aa9d-18181a455731/disk-0.vmdk. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 975.226620] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-085894ad-0cc2-4a23-a489-8f2429652a74 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.234277] env[61970]: DEBUG oslo_vmware.rw_handles [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522eedc8-b148-74f5-aa9d-18181a455731/disk-0.vmdk is in state: ready. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 975.234494] env[61970]: ERROR oslo_vmware.rw_handles [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522eedc8-b148-74f5-aa9d-18181a455731/disk-0.vmdk due to incomplete transfer. [ 975.234779] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f3b6b761-f6ed-4d11-ad2e-052628f72589 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.242327] env[61970]: DEBUG oslo_vmware.rw_handles [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522eedc8-b148-74f5-aa9d-18181a455731/disk-0.vmdk. {{(pid=61970) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 975.242568] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Uploaded image 1a7f3a54-27a5-49f0-9a72-74cdc8f4f51b to the Glance image server {{(pid=61970) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 975.244575] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Destroying the VM {{(pid=61970) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 975.244886] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b6e55110-0f95-4668-9a9f-9815d05a3e05 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.251645] env[61970]: DEBUG oslo_vmware.api [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 975.251645] env[61970]: value = "task-1355966" [ 975.251645] env[61970]: _type = "Task" [ 975.251645] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.261103] env[61970]: DEBUG oslo_vmware.api [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355966, 'name': Destroy_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.381417] env[61970]: DEBUG oslo_concurrency.lockutils [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.314s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.442842] env[61970]: DEBUG oslo_vmware.api [None req-2fbafc6e-bf49-4621-8213-5347c9c3a63d tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355964, 'name': SuspendVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.597462] env[61970]: DEBUG nova.network.neutron [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Updating instance_info_cache with network_info: [{"id": "7bfd07fd-97bc-42b5-9571-0807d864f9bc", "address": "fa:16:3e:89:e6:54", "network": {"id": "42def9ae-4eed-41e5-a538-81eaced2cdd4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2047751481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19dae2316adb4dc980bd972c8c40c034", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c1b8689-a9b4-4972-beb9-6a1c8de1dc88", "external-id": "nsx-vlan-transportzone-455", "segmentation_id": 455, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bfd07fd-97", "ovs_interfaceid": "7bfd07fd-97bc-42b5-9571-0807d864f9bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.637022] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355965, 'name': CreateVM_Task, 'duration_secs': 0.408722} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.637022] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 975.637022] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.637022] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.637022] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 975.637022] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1fbcf3b-57ec-4338-ad09-d3132eeb8d00 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.641761] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 975.641761] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d02a01-4f07-9e68-df02-8259498e70cc" [ 975.641761] env[61970]: _type = "Task" [ 975.641761] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.651290] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d02a01-4f07-9e68-df02-8259498e70cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.700255] env[61970]: DEBUG oslo_vmware.api [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355963, 'name': PowerOnVM_Task, 'duration_secs': 0.543238} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.700816] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 975.701174] env[61970]: INFO nova.compute.manager [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Took 7.84 seconds to spawn the instance on the hypervisor. [ 975.701504] env[61970]: DEBUG nova.compute.manager [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.702421] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e848fc-29f2-4526-a1bf-b5dce2a65ea6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.720296] env[61970]: DEBUG nova.network.neutron [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Successfully updated port: a10357cb-17f1-4e77-ae7f-ac439a804c9c {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 975.761542] env[61970]: DEBUG oslo_vmware.api [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355966, 'name': Destroy_Task, 'duration_secs': 0.370421} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.761958] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Destroyed the VM [ 975.762043] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Deleting Snapshot of the VM instance {{(pid=61970) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 975.762285] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f2b9b45b-fe06-422a-b52f-9568cb9b0134 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.767485] env[61970]: DEBUG oslo_vmware.api [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 975.767485] env[61970]: value = "task-1355967" [ 975.767485] env[61970]: _type = "Task" [ 975.767485] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.774710] env[61970]: DEBUG oslo_vmware.api [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355967, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.841477] env[61970]: DEBUG nova.network.neutron [req-67de96e4-6464-44a9-9873-cfdb4b4cf213 req-3afb3cd8-7cfd-4129-8a16-328d5b1cb894 service nova] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Updated VIF entry in instance network info cache for port 4e65fd7e-c508-4d66-a854-d26fd273014a. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 975.842008] env[61970]: DEBUG nova.network.neutron [req-67de96e4-6464-44a9-9873-cfdb4b4cf213 req-3afb3cd8-7cfd-4129-8a16-328d5b1cb894 service nova] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Updating instance_info_cache with network_info: [{"id": "4e65fd7e-c508-4d66-a854-d26fd273014a", "address": "fa:16:3e:93:cf:a0", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e65fd7e-c5", "ovs_interfaceid": "4e65fd7e-c508-4d66-a854-d26fd273014a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.889344] env[61970]: DEBUG oslo_concurrency.lockutils [None req-79901355-382d-4e9d-ae46-4995eabf79e1 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "7f59000a-94d9-45b6-aa7a-300d95793615" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.138s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.890256] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "7f59000a-94d9-45b6-aa7a-300d95793615" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.839s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.890256] env[61970]: INFO nova.compute.manager [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Unshelving [ 975.940190] env[61970]: DEBUG oslo_vmware.api [None req-2fbafc6e-bf49-4621-8213-5347c9c3a63d tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355964, 'name': SuspendVM_Task, 'duration_secs': 0.634357} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.940512] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2fbafc6e-bf49-4621-8213-5347c9c3a63d tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Suspended the VM {{(pid=61970) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 975.940770] env[61970]: DEBUG nova.compute.manager [None req-2fbafc6e-bf49-4621-8213-5347c9c3a63d tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.942139] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd909cb6-c536-4e20-bc92-dab2cf92569e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.999343] env[61970]: DEBUG nova.compute.manager [req-ced8012c-7a04-475b-9b9c-d364369ae691 req-d867ffd1-9585-453f-8469-5405275639d1 service nova] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Received event network-vif-plugged-a10357cb-17f1-4e77-ae7f-ac439a804c9c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 975.999575] env[61970]: DEBUG oslo_concurrency.lockutils [req-ced8012c-7a04-475b-9b9c-d364369ae691 req-d867ffd1-9585-453f-8469-5405275639d1 service nova] Acquiring lock "3b1e8446-7862-4532-b2ed-611b63dbf391-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.999794] env[61970]: DEBUG oslo_concurrency.lockutils [req-ced8012c-7a04-475b-9b9c-d364369ae691 req-d867ffd1-9585-453f-8469-5405275639d1 service nova] Lock "3b1e8446-7862-4532-b2ed-611b63dbf391-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.999965] env[61970]: DEBUG oslo_concurrency.lockutils [req-ced8012c-7a04-475b-9b9c-d364369ae691 req-d867ffd1-9585-453f-8469-5405275639d1 service nova] Lock "3b1e8446-7862-4532-b2ed-611b63dbf391-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.000184] env[61970]: DEBUG nova.compute.manager [req-ced8012c-7a04-475b-9b9c-d364369ae691 req-d867ffd1-9585-453f-8469-5405275639d1 service nova] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] No waiting events found dispatching network-vif-plugged-a10357cb-17f1-4e77-ae7f-ac439a804c9c {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 976.000361] env[61970]: WARNING nova.compute.manager [req-ced8012c-7a04-475b-9b9c-d364369ae691 req-d867ffd1-9585-453f-8469-5405275639d1 service nova] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Received unexpected event network-vif-plugged-a10357cb-17f1-4e77-ae7f-ac439a804c9c for instance with vm_state building and task_state spawning. [ 976.000714] env[61970]: DEBUG nova.compute.manager [req-ced8012c-7a04-475b-9b9c-d364369ae691 req-d867ffd1-9585-453f-8469-5405275639d1 service nova] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Received event network-changed-a10357cb-17f1-4e77-ae7f-ac439a804c9c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.000714] env[61970]: DEBUG nova.compute.manager [req-ced8012c-7a04-475b-9b9c-d364369ae691 req-d867ffd1-9585-453f-8469-5405275639d1 service nova] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Refreshing instance network info cache due to event network-changed-a10357cb-17f1-4e77-ae7f-ac439a804c9c. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 976.000864] env[61970]: DEBUG oslo_concurrency.lockutils [req-ced8012c-7a04-475b-9b9c-d364369ae691 req-d867ffd1-9585-453f-8469-5405275639d1 service nova] Acquiring lock "refresh_cache-3b1e8446-7862-4532-b2ed-611b63dbf391" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.001010] env[61970]: DEBUG oslo_concurrency.lockutils [req-ced8012c-7a04-475b-9b9c-d364369ae691 req-d867ffd1-9585-453f-8469-5405275639d1 service nova] Acquired lock "refresh_cache-3b1e8446-7862-4532-b2ed-611b63dbf391" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.001257] env[61970]: DEBUG nova.network.neutron [req-ced8012c-7a04-475b-9b9c-d364369ae691 req-d867ffd1-9585-453f-8469-5405275639d1 service nova] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Refreshing network info cache for port a10357cb-17f1-4e77-ae7f-ac439a804c9c {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 976.099173] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Releasing lock "refresh_cache-15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.099470] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Updated the network info_cache for instance {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 976.099470] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 976.099606] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 976.099751] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 976.099902] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 976.100422] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 976.100422] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 976.100422] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61970) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 976.100568] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 976.152873] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d02a01-4f07-9e68-df02-8259498e70cc, 'name': SearchDatastore_Task, 'duration_secs': 0.009731} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.153210] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.153448] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 976.153681] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.153833] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.154130] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 976.154396] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d000742a-34cf-46a9-8f0b-235130e1ea6e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.162233] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 976.162438] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 976.163158] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5b278aa-ac09-486d-957e-0b526d9709bf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.168095] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 976.168095] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528cdc2c-a3df-c761-80bc-788ed798d358" [ 976.168095] env[61970]: _type = "Task" [ 976.168095] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.175329] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528cdc2c-a3df-c761-80bc-788ed798d358, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.221531] env[61970]: DEBUG oslo_concurrency.lockutils [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "refresh_cache-3b1e8446-7862-4532-b2ed-611b63dbf391" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.223144] env[61970]: INFO nova.compute.manager [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Took 14.43 seconds to build instance. [ 976.281401] env[61970]: DEBUG oslo_vmware.api [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355967, 'name': RemoveSnapshot_Task, 'duration_secs': 0.508764} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.281853] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Deleted Snapshot of the VM instance {{(pid=61970) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 976.282293] env[61970]: INFO nova.compute.manager [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Took 13.66 seconds to snapshot the instance on the hypervisor. [ 976.344501] env[61970]: DEBUG oslo_concurrency.lockutils [req-67de96e4-6464-44a9-9873-cfdb4b4cf213 req-3afb3cd8-7cfd-4129-8a16-328d5b1cb894 service nova] Releasing lock "refresh_cache-bc948a41-1a22-46fa-a76a-50b4c4afa4b7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.532255] env[61970]: DEBUG nova.network.neutron [req-ced8012c-7a04-475b-9b9c-d364369ae691 req-d867ffd1-9585-453f-8469-5405275639d1 service nova] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 976.603147] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.603382] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.603545] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.603695] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61970) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 976.604654] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d936bc80-5a6d-4784-9256-66de23debece {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.609621] env[61970]: DEBUG nova.network.neutron [req-ced8012c-7a04-475b-9b9c-d364369ae691 req-d867ffd1-9585-453f-8469-5405275639d1 service nova] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.614055] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0edfa65-220b-4922-a0fa-06e651042e0f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.628727] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa95395-6708-4d43-a201-a2be694aedf2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.635420] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11930ee0-afc3-44e7-9db0-b054b6a911f8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.663529] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180062MB free_disk=150GB free_vcpus=48 pci_devices=None {{(pid=61970) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 976.663649] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.663810] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.676356] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528cdc2c-a3df-c761-80bc-788ed798d358, 'name': SearchDatastore_Task, 'duration_secs': 0.008673} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.677098] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-129d9041-3d1f-4f1e-8d85-99cb8e83848e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.682470] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 976.682470] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d60580-a4aa-8b62-3611-574072e87499" [ 976.682470] env[61970]: _type = "Task" [ 976.682470] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.689153] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d60580-a4aa-8b62-3611-574072e87499, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.724430] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddf2a794-69b8-4b38-bec1-d18b6f0f93e0 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "4d093692-07d3-49aa-8b0a-fb52bd9c0ffc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.942s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.840632] env[61970]: DEBUG nova.compute.manager [None req-fc65301d-80e5-43f5-bb72-ab7befa6d180 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Found 1 images (rotation: 2) {{(pid=61970) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 976.855274] env[61970]: DEBUG oslo_concurrency.lockutils [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "4d093692-07d3-49aa-8b0a-fb52bd9c0ffc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.855524] env[61970]: DEBUG oslo_concurrency.lockutils [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "4d093692-07d3-49aa-8b0a-fb52bd9c0ffc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.855727] env[61970]: DEBUG oslo_concurrency.lockutils [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "4d093692-07d3-49aa-8b0a-fb52bd9c0ffc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.855906] env[61970]: DEBUG oslo_concurrency.lockutils [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "4d093692-07d3-49aa-8b0a-fb52bd9c0ffc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.856086] env[61970]: DEBUG oslo_concurrency.lockutils [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "4d093692-07d3-49aa-8b0a-fb52bd9c0ffc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.857986] env[61970]: INFO nova.compute.manager [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Terminating instance [ 976.859745] env[61970]: DEBUG nova.compute.manager [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 976.859940] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 976.861116] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f1a946-ee1e-4961-b1e2-30c3cf6c0a84 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.872320] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 976.872575] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9135fa3c-f16f-4820-875d-4db2d4764ea3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.879051] env[61970]: DEBUG oslo_vmware.api [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 976.879051] env[61970]: value = "task-1355968" [ 976.879051] env[61970]: _type = "Task" [ 976.879051] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.887417] env[61970]: DEBUG oslo_vmware.api [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355968, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.898653] env[61970]: DEBUG nova.compute.utils [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 977.114933] env[61970]: DEBUG oslo_concurrency.lockutils [req-ced8012c-7a04-475b-9b9c-d364369ae691 req-d867ffd1-9585-453f-8469-5405275639d1 service nova] Releasing lock "refresh_cache-3b1e8446-7862-4532-b2ed-611b63dbf391" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.115303] env[61970]: DEBUG oslo_concurrency.lockutils [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "refresh_cache-3b1e8446-7862-4532-b2ed-611b63dbf391" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.115303] env[61970]: DEBUG nova.network.neutron [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 977.192599] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d60580-a4aa-8b62-3611-574072e87499, 'name': SearchDatastore_Task, 'duration_secs': 0.009034} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.192869] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.193145] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] bc948a41-1a22-46fa-a76a-50b4c4afa4b7/bc948a41-1a22-46fa-a76a-50b4c4afa4b7.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 977.193413] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-898914e5-f348-4931-b72b-103616fc03e4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.200226] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 977.200226] env[61970]: value = "task-1355969" [ 977.200226] env[61970]: _type = "Task" [ 977.200226] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.207993] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355969, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.368917] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "479342e0-7627-4dda-a589-a7bc5a5c4270" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.369192] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "479342e0-7627-4dda-a589-a7bc5a5c4270" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.369430] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "479342e0-7627-4dda-a589-a7bc5a5c4270-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.369659] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "479342e0-7627-4dda-a589-a7bc5a5c4270-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.369968] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "479342e0-7627-4dda-a589-a7bc5a5c4270-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.372334] env[61970]: INFO nova.compute.manager [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Terminating instance [ 977.374512] env[61970]: DEBUG nova.compute.manager [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 977.374715] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 977.375616] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b310d980-2e83-48a3-b81b-de757e98cf13 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.385323] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 977.386033] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1f97cf96-37af-4c96-ab7e-4bcd6bce16c9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.390833] env[61970]: DEBUG oslo_vmware.api [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355968, 'name': PowerOffVM_Task, 'duration_secs': 0.385972} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.391625] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 977.391782] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 977.392090] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d75caccf-d9a6-4e7f-842e-6de67c24ae9b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.402797] env[61970]: INFO nova.virt.block_device [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Booting with volume a78067a3-226b-4aec-bb5b-6fb37d5bfca3 at /dev/sdb [ 977.443974] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f3e74f6f-4a55-4539-a181-e38934b317e5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.455156] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a29e902-6e1b-485c-94db-f01b7ddb96e2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.475574] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 977.475811] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 977.476018] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Deleting the datastore file [datastore1] 479342e0-7627-4dda-a589-a7bc5a5c4270 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 977.476357] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4fd67fe-d398-489b-84d1-2498dbcfb620 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.489140] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0c6f3d12-9fbc-4ed8-a77e-4542fdd8fb32 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.491468] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 977.491468] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 977.491697] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleting the datastore file [datastore2] 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 977.493233] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-38555336-21db-4b0f-a7b1-657149b1d610 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.495230] env[61970]: DEBUG oslo_vmware.api [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 977.495230] env[61970]: value = "task-1355972" [ 977.495230] env[61970]: _type = "Task" [ 977.495230] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.501759] env[61970]: DEBUG oslo_vmware.api [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 977.501759] env[61970]: value = "task-1355973" [ 977.501759] env[61970]: _type = "Task" [ 977.501759] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.505162] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f07c10e-c42f-4448-9775-9de62cf99067 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.523621] env[61970]: DEBUG oslo_vmware.api [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355972, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.530793] env[61970]: DEBUG oslo_vmware.api [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355973, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.540466] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413bd567-5c5e-4fd6-bbdb-929ef95f9557 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.546296] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed70c39b-076c-49eb-8f66-0a28b48838b4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.559439] env[61970]: DEBUG nova.virt.block_device [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Updating existing volume attachment record: eedbcc21-7f74-484a-94dd-1a6c0645a263 {{(pid=61970) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 977.648604] env[61970]: DEBUG nova.network.neutron [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 977.693999] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.694180] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 0304640a-6c46-4f57-a8b8-5058da2e3de4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.694307] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance d0ad9689-bdb7-4d68-b19a-3f805ab4612e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.694491] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance ebb92a5b-f635-4504-a57f-395de514015a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.694578] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.694632] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance c4ec1797-d939-4c24-9314-ce9c74eb5aa1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.694744] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.694854] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 479342e0-7627-4dda-a589-a7bc5a5c4270 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.694963] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.695083] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance bc948a41-1a22-46fa-a76a-50b4c4afa4b7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.695195] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 3b1e8446-7862-4532-b2ed-611b63dbf391 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.709593] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355969, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466891} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.709848] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] bc948a41-1a22-46fa-a76a-50b4c4afa4b7/bc948a41-1a22-46fa-a76a-50b4c4afa4b7.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 977.710086] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 977.710340] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dc857fe7-9701-46bc-80f3-deffdc2d96ca {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.716884] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 977.716884] env[61970]: value = "task-1355975" [ 977.716884] env[61970]: _type = "Task" [ 977.716884] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.724642] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355975, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.802478] env[61970]: DEBUG nova.network.neutron [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Updating instance_info_cache with network_info: [{"id": "a10357cb-17f1-4e77-ae7f-ac439a804c9c", "address": "fa:16:3e:ac:8e:45", "network": {"id": "cba185ff-bfa3-4547-b7b0-638c3e79a24c", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-864852491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5c64cb9232fb413cbd7627dcf077e9ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa10357cb-17", "ovs_interfaceid": "a10357cb-17f1-4e77-ae7f-ac439a804c9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.830802] env[61970]: DEBUG nova.compute.manager [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 977.831712] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c159f4-fcad-4539-9cf3-7bf78f879588 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.006064] env[61970]: DEBUG oslo_vmware.api [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1355972, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.261658} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.007026] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 978.007026] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 978.007026] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 978.007026] env[61970]: INFO nova.compute.manager [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Took 0.63 seconds to destroy the instance on the hypervisor. [ 978.007265] env[61970]: DEBUG oslo.service.loopingcall [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.007432] env[61970]: DEBUG nova.compute.manager [-] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 978.007527] env[61970]: DEBUG nova.network.neutron [-] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 978.027978] env[61970]: DEBUG oslo_vmware.api [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1355973, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.278627} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.027978] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 978.028762] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 978.028762] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 978.028762] env[61970]: INFO nova.compute.manager [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Took 1.17 seconds to destroy the instance on the hypervisor. [ 978.028949] env[61970]: DEBUG oslo.service.loopingcall [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.029181] env[61970]: DEBUG nova.compute.manager [-] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 978.029261] env[61970]: DEBUG nova.network.neutron [-] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 978.202679] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 7f59000a-94d9-45b6-aa7a-300d95793615 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 978.202679] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 978.202679] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2624MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 978.230728] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355975, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062735} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.233413] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 978.234471] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb931513-2868-4058-8844-50cae66306bd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.257394] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] bc948a41-1a22-46fa-a76a-50b4c4afa4b7/bc948a41-1a22-46fa-a76a-50b4c4afa4b7.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 978.260248] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24a642a0-0ea4-4e9d-a3f2-19b4d359986c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.284142] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 978.284142] env[61970]: value = "task-1355978" [ 978.284142] env[61970]: _type = "Task" [ 978.284142] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.292744] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355978, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.305319] env[61970]: DEBUG oslo_concurrency.lockutils [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "refresh_cache-3b1e8446-7862-4532-b2ed-611b63dbf391" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.305625] env[61970]: DEBUG nova.compute.manager [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Instance network_info: |[{"id": "a10357cb-17f1-4e77-ae7f-ac439a804c9c", "address": "fa:16:3e:ac:8e:45", "network": {"id": "cba185ff-bfa3-4547-b7b0-638c3e79a24c", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-864852491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5c64cb9232fb413cbd7627dcf077e9ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa10357cb-17", "ovs_interfaceid": "a10357cb-17f1-4e77-ae7f-ac439a804c9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 978.306051] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:8e:45', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '40c947c4-f471-4d48-8e43-fee54198107e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a10357cb-17f1-4e77-ae7f-ac439a804c9c', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 978.314121] env[61970]: DEBUG oslo.service.loopingcall [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.315228] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 978.315467] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7052d5b8-14e4-45f9-b62d-d907d1cf65c4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.334651] env[61970]: DEBUG nova.compute.manager [req-2171d6fd-0a06-4df4-915c-9d723296ab45 req-145f63ed-3175-40c6-8e91-d3591cf975bb service nova] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Received event network-vif-deleted-f73342d3-c298-4151-9bba-79202180cd9a {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.334855] env[61970]: INFO nova.compute.manager [req-2171d6fd-0a06-4df4-915c-9d723296ab45 req-145f63ed-3175-40c6-8e91-d3591cf975bb service nova] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Neutron deleted interface f73342d3-c298-4151-9bba-79202180cd9a; detaching it from the instance and deleting it from the info cache [ 978.335055] env[61970]: DEBUG nova.network.neutron [req-2171d6fd-0a06-4df4-915c-9d723296ab45 req-145f63ed-3175-40c6-8e91-d3591cf975bb service nova] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.336700] env[61970]: DEBUG nova.compute.manager [req-aec1cf6e-d38a-4ff5-85c3-4b191a5d0130 req-e2ca7b95-7664-469b-8abb-cacce8373e40 service nova] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Received event network-vif-deleted-eb47f433-6083-4086-ac69-1df8051e0d2f {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.336889] env[61970]: INFO nova.compute.manager [req-aec1cf6e-d38a-4ff5-85c3-4b191a5d0130 req-e2ca7b95-7664-469b-8abb-cacce8373e40 service nova] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Neutron deleted interface eb47f433-6083-4086-ac69-1df8051e0d2f; detaching it from the instance and deleting it from the info cache [ 978.337067] env[61970]: DEBUG nova.network.neutron [req-aec1cf6e-d38a-4ff5-85c3-4b191a5d0130 req-e2ca7b95-7664-469b-8abb-cacce8373e40 service nova] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.341726] env[61970]: INFO nova.compute.manager [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] instance snapshotting [ 978.342328] env[61970]: DEBUG nova.objects.instance [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lazy-loading 'flavor' on Instance uuid ebb92a5b-f635-4504-a57f-395de514015a {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.346657] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 978.346657] env[61970]: value = "task-1355979" [ 978.346657] env[61970]: _type = "Task" [ 978.346657] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.358615] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355979, 'name': CreateVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.435860] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db3fd8f-8470-48b3-b568-0ceeaeabefb2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.443079] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27cedf2-0d5d-405a-818a-b5b9c691295e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.473489] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e00b985-c200-4ad4-b183-b0cf195144ea {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.480700] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72458db7-db6a-42cf-9685-2618258eabb8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.493595] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.793720] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355978, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.802156] env[61970]: DEBUG nova.network.neutron [-] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.803798] env[61970]: DEBUG nova.network.neutron [-] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.839933] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2f693c17-50e3-4961-a7c5-77b91aa36235 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.842560] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-003abb7b-6f8e-460f-97be-6f93ccb34ae5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.857646] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0994515-4e60-4761-bf51-f05316f041b2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.871048] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f427c44e-3a76-45b0-bf88-bf187c49d7a4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.885242] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-065dc724-68b4-4872-b71a-bf74f3fdbd1e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.893684] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355979, 'name': CreateVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.907218] env[61970]: DEBUG nova.compute.manager [req-aec1cf6e-d38a-4ff5-85c3-4b191a5d0130 req-e2ca7b95-7664-469b-8abb-cacce8373e40 service nova] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Detach interface failed, port_id=eb47f433-6083-4086-ac69-1df8051e0d2f, reason: Instance 479342e0-7627-4dda-a589-a7bc5a5c4270 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 978.927710] env[61970]: DEBUG nova.compute.manager [req-2171d6fd-0a06-4df4-915c-9d723296ab45 req-145f63ed-3175-40c6-8e91-d3591cf975bb service nova] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Detach interface failed, port_id=f73342d3-c298-4151-9bba-79202180cd9a, reason: Instance 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 978.928821] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-190aa914-3633-48c1-9114-f412d4c3865e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.997349] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 979.177675] env[61970]: DEBUG oslo_concurrency.lockutils [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.177973] env[61970]: DEBUG oslo_concurrency.lockutils [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.178239] env[61970]: DEBUG oslo_concurrency.lockutils [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.178834] env[61970]: DEBUG oslo_concurrency.lockutils [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.179035] env[61970]: DEBUG oslo_concurrency.lockutils [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.181020] env[61970]: INFO nova.compute.manager [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Terminating instance [ 979.182823] env[61970]: DEBUG nova.compute.manager [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 979.183035] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 979.183857] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9ab14f-4de8-495d-8ec9-eb19fec7571a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.191097] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 979.191299] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ebdafec-2faa-4e46-8594-e41a5057a85c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.196855] env[61970]: DEBUG oslo_vmware.api [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 979.196855] env[61970]: value = "task-1355980" [ 979.196855] env[61970]: _type = "Task" [ 979.196855] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.204027] env[61970]: DEBUG oslo_vmware.api [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355980, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.294980] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355978, 'name': ReconfigVM_Task, 'duration_secs': 0.680167} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.295258] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Reconfigured VM instance instance-0000005d to attach disk [datastore2] bc948a41-1a22-46fa-a76a-50b4c4afa4b7/bc948a41-1a22-46fa-a76a-50b4c4afa4b7.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 979.295939] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-96f69d19-52df-4595-8b91-c9f5e25713e7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.302324] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 979.302324] env[61970]: value = "task-1355981" [ 979.302324] env[61970]: _type = "Task" [ 979.302324] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.305519] env[61970]: INFO nova.compute.manager [-] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Took 1.30 seconds to deallocate network for instance. [ 979.305835] env[61970]: INFO nova.compute.manager [-] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Took 1.28 seconds to deallocate network for instance. [ 979.317381] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355981, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.363560] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1355979, 'name': CreateVM_Task, 'duration_secs': 0.56323} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.363738] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 979.364422] env[61970]: DEBUG oslo_concurrency.lockutils [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.364594] env[61970]: DEBUG oslo_concurrency.lockutils [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.364921] env[61970]: DEBUG oslo_concurrency.lockutils [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 979.365211] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f817421f-dff1-4e7f-a150-55b5ca17d2a6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.370551] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 979.370551] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52783a7d-05a5-4615-f7b0-2ec4c7b0b045" [ 979.370551] env[61970]: _type = "Task" [ 979.370551] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.378729] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52783a7d-05a5-4615-f7b0-2ec4c7b0b045, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.440660] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Creating Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 979.440986] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-22e02c2e-40f9-4a7e-bb3c-8f78f8e26a0d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.447511] env[61970]: DEBUG oslo_vmware.api [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 979.447511] env[61970]: value = "task-1355982" [ 979.447511] env[61970]: _type = "Task" [ 979.447511] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.454943] env[61970]: DEBUG oslo_vmware.api [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355982, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.502521] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 979.502803] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.839s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.706639] env[61970]: DEBUG oslo_vmware.api [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355980, 'name': PowerOffVM_Task, 'duration_secs': 0.310124} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.706639] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 979.706887] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 979.707040] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-57766dc9-0dcc-4469-8052-3063351566ab {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.768676] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 979.768910] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 979.769110] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Deleting the datastore file [datastore2] ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 979.769604] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5663e490-c4cd-45e0-9cd4-fa09ea53e2f0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.775649] env[61970]: DEBUG oslo_vmware.api [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 979.775649] env[61970]: value = "task-1355984" [ 979.775649] env[61970]: _type = "Task" [ 979.775649] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.783208] env[61970]: DEBUG oslo_vmware.api [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355984, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.811712] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355981, 'name': Rename_Task, 'duration_secs': 0.17499} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.811990] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 979.812255] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-14a52568-0fa7-4bb9-b7e2-b99a82be0c65 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.816964] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 979.816964] env[61970]: value = "task-1355986" [ 979.816964] env[61970]: _type = "Task" [ 979.816964] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.820611] env[61970]: DEBUG oslo_concurrency.lockutils [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.820871] env[61970]: DEBUG oslo_concurrency.lockutils [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.821126] env[61970]: DEBUG nova.objects.instance [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lazy-loading 'resources' on Instance uuid 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.822916] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.827759] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355986, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.880293] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52783a7d-05a5-4615-f7b0-2ec4c7b0b045, 'name': SearchDatastore_Task, 'duration_secs': 0.010011} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.880678] env[61970]: DEBUG oslo_concurrency.lockutils [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.880959] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 979.881250] env[61970]: DEBUG oslo_concurrency.lockutils [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.881425] env[61970]: DEBUG oslo_concurrency.lockutils [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.881643] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 979.881943] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4712ae85-6b55-4712-92f8-a6b18e555389 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.889259] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 979.889451] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 979.890470] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6c4554e-a388-4cf2-9fdd-2449d4957c28 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.895103] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 979.895103] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]521310ed-fb71-3d66-112e-503fcf233fb0" [ 979.895103] env[61970]: _type = "Task" [ 979.895103] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.902569] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]521310ed-fb71-3d66-112e-503fcf233fb0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.956760] env[61970]: DEBUG oslo_vmware.api [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355982, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.285505] env[61970]: DEBUG oslo_vmware.api [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355984, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133924} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.285872] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 980.285935] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 980.286121] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 980.286302] env[61970]: INFO nova.compute.manager [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Took 1.10 seconds to destroy the instance on the hypervisor. [ 980.286544] env[61970]: DEBUG oslo.service.loopingcall [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 980.286743] env[61970]: DEBUG nova.compute.manager [-] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 980.286838] env[61970]: DEBUG nova.network.neutron [-] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 980.329469] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355986, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.404848] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]521310ed-fb71-3d66-112e-503fcf233fb0, 'name': SearchDatastore_Task, 'duration_secs': 0.012128} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.408193] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27efbaf5-8691-40dc-805a-ecca19181465 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.416228] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 980.416228] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52274a33-ac6d-ef1a-5333-498767cdab63" [ 980.416228] env[61970]: _type = "Task" [ 980.416228] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.430524] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52274a33-ac6d-ef1a-5333-498767cdab63, 'name': SearchDatastore_Task, 'duration_secs': 0.00887} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.430967] env[61970]: DEBUG oslo_concurrency.lockutils [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.431248] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 3b1e8446-7862-4532-b2ed-611b63dbf391/3b1e8446-7862-4532-b2ed-611b63dbf391.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 980.431502] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b695cfe2-c498-4c76-a3d1-10796eeccf61 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.437773] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 980.437773] env[61970]: value = "task-1355987" [ 980.437773] env[61970]: _type = "Task" [ 980.437773] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.448651] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355987, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.460505] env[61970]: DEBUG oslo_vmware.api [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355982, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.508851] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d57fc2e4-37b2-41fa-91af-0b12405b15f1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.517675] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f03f496f-f821-4d52-9937-da6c709f8300 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.551239] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c928ec0-bcb2-49dc-9d11-161765dd5fbe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.558046] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db80f6cf-b461-4410-8cb5-19c787ffca78 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.572058] env[61970]: DEBUG nova.compute.provider_tree [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.810219] env[61970]: DEBUG nova.compute.manager [req-c8c66070-5847-4780-8a75-5b442c599146 req-52e354d3-9be1-45ab-a271-18da86771798 service nova] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Received event network-vif-deleted-28a89dd6-da27-49dd-a87c-a702154014c1 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.810219] env[61970]: INFO nova.compute.manager [req-c8c66070-5847-4780-8a75-5b442c599146 req-52e354d3-9be1-45ab-a271-18da86771798 service nova] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Neutron deleted interface 28a89dd6-da27-49dd-a87c-a702154014c1; detaching it from the instance and deleting it from the info cache [ 980.810219] env[61970]: DEBUG nova.network.neutron [req-c8c66070-5847-4780-8a75-5b442c599146 req-52e354d3-9be1-45ab-a271-18da86771798 service nova] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.829810] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355986, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.947485] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355987, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.495616} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.947818] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 3b1e8446-7862-4532-b2ed-611b63dbf391/3b1e8446-7862-4532-b2ed-611b63dbf391.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 980.948047] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 980.948329] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-43b06465-043e-44f1-bcdc-9c4bc8dc8431 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.960747] env[61970]: DEBUG oslo_vmware.api [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355982, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.962377] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 980.962377] env[61970]: value = "task-1355988" [ 980.962377] env[61970]: _type = "Task" [ 980.962377] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.970665] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355988, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.075943] env[61970]: DEBUG nova.scheduler.client.report [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 981.282146] env[61970]: DEBUG nova.network.neutron [-] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.313398] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-caf7c98b-d9d4-4801-914f-c6dae0ae6508 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.326819] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ab5e5a-ffa2-402c-8c79-c198cc4be697 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.340227] env[61970]: DEBUG oslo_vmware.api [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1355986, 'name': PowerOnVM_Task, 'duration_secs': 1.3834} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.340557] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 981.340769] env[61970]: INFO nova.compute.manager [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Took 8.95 seconds to spawn the instance on the hypervisor. [ 981.340975] env[61970]: DEBUG nova.compute.manager [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 981.341751] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c999c2-5941-4018-b3c6-e6b7064fc777 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.357654] env[61970]: DEBUG nova.compute.manager [req-c8c66070-5847-4780-8a75-5b442c599146 req-52e354d3-9be1-45ab-a271-18da86771798 service nova] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Detach interface failed, port_id=28a89dd6-da27-49dd-a87c-a702154014c1, reason: Instance ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 981.460375] env[61970]: DEBUG oslo_vmware.api [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355982, 'name': CreateSnapshot_Task, 'duration_secs': 1.780883} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.460575] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Created Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 981.461355] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c3b6d0-b3c3-4342-9f53-dbbedb18c47b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.471442] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355988, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072619} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.474440] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 981.477258] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a27c06e2-1e8d-4f82-92e5-3d2876c4cc2a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.500258] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 3b1e8446-7862-4532-b2ed-611b63dbf391/3b1e8446-7862-4532-b2ed-611b63dbf391.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 981.500519] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4831c7c3-3028-4e24-be3b-5d8a85e66833 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.520485] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 981.520485] env[61970]: value = "task-1355989" [ 981.520485] env[61970]: _type = "Task" [ 981.520485] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.528446] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355989, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.581366] env[61970]: DEBUG oslo_concurrency.lockutils [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.759s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.583128] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.760s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.583952] env[61970]: DEBUG nova.objects.instance [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lazy-loading 'resources' on Instance uuid 479342e0-7627-4dda-a589-a7bc5a5c4270 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.602115] env[61970]: INFO nova.scheduler.client.report [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleted allocations for instance 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc [ 981.785115] env[61970]: INFO nova.compute.manager [-] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Took 1.50 seconds to deallocate network for instance. [ 981.789792] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "interface-c4ec1797-d939-4c24-9314-ce9c74eb5aa1-90fbe562-a7de-4ce8-a337-3c3f7be95120" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.790111] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "interface-c4ec1797-d939-4c24-9314-ce9c74eb5aa1-90fbe562-a7de-4ce8-a337-3c3f7be95120" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.790511] env[61970]: DEBUG nova.objects.instance [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lazy-loading 'flavor' on Instance uuid c4ec1797-d939-4c24-9314-ce9c74eb5aa1 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.867552] env[61970]: INFO nova.compute.manager [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Took 16.67 seconds to build instance. [ 981.989941] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Creating linked-clone VM from snapshot {{(pid=61970) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 981.990280] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-75c0347c-8a73-4c02-a82c-72d4e45a5edf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.999039] env[61970]: DEBUG oslo_vmware.api [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 981.999039] env[61970]: value = "task-1355990" [ 981.999039] env[61970]: _type = "Task" [ 981.999039] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.007812] env[61970]: DEBUG oslo_vmware.api [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355990, 'name': CloneVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.029992] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355989, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.116373] env[61970]: DEBUG oslo_concurrency.lockutils [None req-947eb71f-0118-42ed-a0b6-75330dcd2a65 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "4d093692-07d3-49aa-8b0a-fb52bd9c0ffc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.259s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.247403] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ed9f245-2912-40d8-8227-e7a690773ee6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.255195] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cafcd54-7b6a-4326-ace9-8a0d04451fd8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.288433] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300fd26b-4a96-4af6-909c-f1fb79a75d09 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.293415] env[61970]: DEBUG oslo_concurrency.lockutils [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.300057] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d704e8f-086e-42de-811e-d317a1550a8d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.314273] env[61970]: DEBUG nova.compute.provider_tree [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.370513] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef59f963-7e03-4316-b3ce-c6116608da49 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.181s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.439488] env[61970]: DEBUG nova.objects.instance [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lazy-loading 'pci_requests' on Instance uuid c4ec1797-d939-4c24-9314-ce9c74eb5aa1 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.513410] env[61970]: DEBUG oslo_vmware.api [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355990, 'name': CloneVM_Task} progress is 94%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.530664] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355989, 'name': ReconfigVM_Task, 'duration_secs': 0.790448} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.530954] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 3b1e8446-7862-4532-b2ed-611b63dbf391/3b1e8446-7862-4532-b2ed-611b63dbf391.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 982.531673] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fb8b94c2-f41f-40b7-a04a-a09dd87eec82 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.537202] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 982.537202] env[61970]: value = "task-1355991" [ 982.537202] env[61970]: _type = "Task" [ 982.537202] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.545283] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355991, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.818035] env[61970]: DEBUG nova.scheduler.client.report [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.836853] env[61970]: DEBUG nova.compute.manager [req-715800c2-a4be-4c6c-afa8-ff9a69ae01a0 req-461458b9-cef0-4653-8046-ab049cc699fb service nova] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Received event network-changed-4e65fd7e-c508-4d66-a854-d26fd273014a {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.837071] env[61970]: DEBUG nova.compute.manager [req-715800c2-a4be-4c6c-afa8-ff9a69ae01a0 req-461458b9-cef0-4653-8046-ab049cc699fb service nova] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Refreshing instance network info cache due to event network-changed-4e65fd7e-c508-4d66-a854-d26fd273014a. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 982.837350] env[61970]: DEBUG oslo_concurrency.lockutils [req-715800c2-a4be-4c6c-afa8-ff9a69ae01a0 req-461458b9-cef0-4653-8046-ab049cc699fb service nova] Acquiring lock "refresh_cache-bc948a41-1a22-46fa-a76a-50b4c4afa4b7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.837499] env[61970]: DEBUG oslo_concurrency.lockutils [req-715800c2-a4be-4c6c-afa8-ff9a69ae01a0 req-461458b9-cef0-4653-8046-ab049cc699fb service nova] Acquired lock "refresh_cache-bc948a41-1a22-46fa-a76a-50b4c4afa4b7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.837650] env[61970]: DEBUG nova.network.neutron [req-715800c2-a4be-4c6c-afa8-ff9a69ae01a0 req-461458b9-cef0-4653-8046-ab049cc699fb service nova] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Refreshing network info cache for port 4e65fd7e-c508-4d66-a854-d26fd273014a {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 982.884614] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "00e3287d-24fb-4ea9-b212-54215828d5df" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.884854] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "00e3287d-24fb-4ea9-b212-54215828d5df" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.942482] env[61970]: DEBUG nova.objects.base [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 982.942755] env[61970]: DEBUG nova.network.neutron [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 983.009583] env[61970]: DEBUG oslo_vmware.api [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1355990, 'name': CloneVM_Task, 'duration_secs': 0.824269} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.010984] env[61970]: DEBUG nova.policy [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2af915dc43c84940a937dfb8a04ea0b6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17ca05a1e6664430a02de563d98c1148', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 983.012700] env[61970]: INFO nova.virt.vmwareapi.vmops [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Created linked-clone VM from snapshot [ 983.013447] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2739f6b3-4c8a-4a33-b117-ebcc48e83cc1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.022251] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Uploading image de2205ef-3756-4fc4-896c-06f28059131f {{(pid=61970) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 983.043062] env[61970]: DEBUG oslo_vmware.rw_handles [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 983.043062] env[61970]: value = "vm-288856" [ 983.043062] env[61970]: _type = "VirtualMachine" [ 983.043062] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 983.043327] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-a116d7cf-ba33-4432-9044-3bd4a63e17ad {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.049319] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355991, 'name': Rename_Task, 'duration_secs': 0.166757} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.050594] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 983.050886] env[61970]: DEBUG oslo_vmware.rw_handles [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lease: (returnval){ [ 983.050886] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f6cfab-b6d4-a81b-1c6c-646fdee8564f" [ 983.050886] env[61970]: _type = "HttpNfcLease" [ 983.050886] env[61970]: } obtained for exporting VM: (result){ [ 983.050886] env[61970]: value = "vm-288856" [ 983.050886] env[61970]: _type = "VirtualMachine" [ 983.050886] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 983.051171] env[61970]: DEBUG oslo_vmware.api [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the lease: (returnval){ [ 983.051171] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f6cfab-b6d4-a81b-1c6c-646fdee8564f" [ 983.051171] env[61970]: _type = "HttpNfcLease" [ 983.051171] env[61970]: } to be ready. {{(pid=61970) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 983.051297] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ffcd8f91-d5dd-4330-a219-b79140936b9b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.058984] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 983.058984] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f6cfab-b6d4-a81b-1c6c-646fdee8564f" [ 983.058984] env[61970]: _type = "HttpNfcLease" [ 983.058984] env[61970]: } is initializing. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 983.060214] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 983.060214] env[61970]: value = "task-1355993" [ 983.060214] env[61970]: _type = "Task" [ 983.060214] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.069768] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355993, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.176143] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.323553] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.740s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.325762] env[61970]: DEBUG oslo_concurrency.lockutils [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.032s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.326008] env[61970]: DEBUG nova.objects.instance [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lazy-loading 'resources' on Instance uuid ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.342959] env[61970]: INFO nova.scheduler.client.report [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Deleted allocations for instance 479342e0-7627-4dda-a589-a7bc5a5c4270 [ 983.387348] env[61970]: DEBUG nova.compute.manager [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 983.561571] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 983.561571] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f6cfab-b6d4-a81b-1c6c-646fdee8564f" [ 983.561571] env[61970]: _type = "HttpNfcLease" [ 983.561571] env[61970]: } is ready. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 983.564956] env[61970]: DEBUG oslo_vmware.rw_handles [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 983.564956] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f6cfab-b6d4-a81b-1c6c-646fdee8564f" [ 983.564956] env[61970]: _type = "HttpNfcLease" [ 983.564956] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 983.565667] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea3ca5c-e1c8-470c-8895-a08e36a422a4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.573664] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355993, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.577176] env[61970]: DEBUG oslo_vmware.rw_handles [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525d1550-fb59-aae7-beee-d43377562f35/disk-0.vmdk from lease info. {{(pid=61970) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 983.577396] env[61970]: DEBUG oslo_vmware.rw_handles [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525d1550-fb59-aae7-beee-d43377562f35/disk-0.vmdk for reading. {{(pid=61970) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 983.638933] env[61970]: DEBUG nova.network.neutron [req-715800c2-a4be-4c6c-afa8-ff9a69ae01a0 req-461458b9-cef0-4653-8046-ab049cc699fb service nova] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Updated VIF entry in instance network info cache for port 4e65fd7e-c508-4d66-a854-d26fd273014a. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 983.639361] env[61970]: DEBUG nova.network.neutron [req-715800c2-a4be-4c6c-afa8-ff9a69ae01a0 req-461458b9-cef0-4653-8046-ab049cc699fb service nova] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Updating instance_info_cache with network_info: [{"id": "4e65fd7e-c508-4d66-a854-d26fd273014a", "address": "fa:16:3e:93:cf:a0", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e65fd7e-c5", "ovs_interfaceid": "4e65fd7e-c508-4d66-a854-d26fd273014a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.685308] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9b5e037d-4085-4367-ae91-b35648f68bb7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.849431] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a799e4c3-3496-4bea-9664-a560de9b46b4 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "479342e0-7627-4dda-a589-a7bc5a5c4270" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.480s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.906751] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.999365] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b5f7c65-cfbe-42bb-b5a9-7c10ffc25238 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.008625] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bceb7fab-9151-4bc1-9d61-2a70d14a7269 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.043694] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6172ba05-395d-42fa-a34f-2328aff562e5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.052875] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-107c4b30-52f4-4c50-b604-0eed7e6d07d4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.068942] env[61970]: DEBUG nova.compute.provider_tree [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.078974] env[61970]: DEBUG oslo_vmware.api [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355993, 'name': PowerOnVM_Task, 'duration_secs': 0.731205} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.080087] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 984.080559] env[61970]: INFO nova.compute.manager [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Took 8.99 seconds to spawn the instance on the hypervisor. [ 984.080819] env[61970]: DEBUG nova.compute.manager [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 984.081941] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd7eba5-d9c1-49ca-a698-72c35eb06b1e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.143196] env[61970]: DEBUG oslo_concurrency.lockutils [req-715800c2-a4be-4c6c-afa8-ff9a69ae01a0 req-461458b9-cef0-4653-8046-ab049cc699fb service nova] Releasing lock "refresh_cache-bc948a41-1a22-46fa-a76a-50b4c4afa4b7" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.532463] env[61970]: DEBUG nova.network.neutron [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Successfully updated port: 90fbe562-a7de-4ce8-a337-3c3f7be95120 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 984.575467] env[61970]: DEBUG nova.scheduler.client.report [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.599762] env[61970]: INFO nova.compute.manager [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Took 13.93 seconds to build instance. [ 984.927765] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "d38bde37-994b-4d58-b4b1-5f79b59fd2aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.928153] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "d38bde37-994b-4d58-b4b1-5f79b59fd2aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.941065] env[61970]: DEBUG nova.compute.manager [req-21133f13-1e98-4ec3-8433-50925a744921 req-73c5976d-c111-4b80-8f7e-fc6e58adabc0 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Received event network-vif-plugged-90fbe562-a7de-4ce8-a337-3c3f7be95120 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.941297] env[61970]: DEBUG oslo_concurrency.lockutils [req-21133f13-1e98-4ec3-8433-50925a744921 req-73c5976d-c111-4b80-8f7e-fc6e58adabc0 service nova] Acquiring lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.941564] env[61970]: DEBUG oslo_concurrency.lockutils [req-21133f13-1e98-4ec3-8433-50925a744921 req-73c5976d-c111-4b80-8f7e-fc6e58adabc0 service nova] Lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.941769] env[61970]: DEBUG oslo_concurrency.lockutils [req-21133f13-1e98-4ec3-8433-50925a744921 req-73c5976d-c111-4b80-8f7e-fc6e58adabc0 service nova] Lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.942011] env[61970]: DEBUG nova.compute.manager [req-21133f13-1e98-4ec3-8433-50925a744921 req-73c5976d-c111-4b80-8f7e-fc6e58adabc0 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] No waiting events found dispatching network-vif-plugged-90fbe562-a7de-4ce8-a337-3c3f7be95120 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 984.942237] env[61970]: WARNING nova.compute.manager [req-21133f13-1e98-4ec3-8433-50925a744921 req-73c5976d-c111-4b80-8f7e-fc6e58adabc0 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Received unexpected event network-vif-plugged-90fbe562-a7de-4ce8-a337-3c3f7be95120 for instance with vm_state active and task_state None. [ 984.942460] env[61970]: DEBUG nova.compute.manager [req-21133f13-1e98-4ec3-8433-50925a744921 req-73c5976d-c111-4b80-8f7e-fc6e58adabc0 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Received event network-changed-90fbe562-a7de-4ce8-a337-3c3f7be95120 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.942749] env[61970]: DEBUG nova.compute.manager [req-21133f13-1e98-4ec3-8433-50925a744921 req-73c5976d-c111-4b80-8f7e-fc6e58adabc0 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Refreshing instance network info cache due to event network-changed-90fbe562-a7de-4ce8-a337-3c3f7be95120. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 984.943078] env[61970]: DEBUG oslo_concurrency.lockutils [req-21133f13-1e98-4ec3-8433-50925a744921 req-73c5976d-c111-4b80-8f7e-fc6e58adabc0 service nova] Acquiring lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.943231] env[61970]: DEBUG oslo_concurrency.lockutils [req-21133f13-1e98-4ec3-8433-50925a744921 req-73c5976d-c111-4b80-8f7e-fc6e58adabc0 service nova] Acquired lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.943494] env[61970]: DEBUG nova.network.neutron [req-21133f13-1e98-4ec3-8433-50925a744921 req-73c5976d-c111-4b80-8f7e-fc6e58adabc0 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Refreshing network info cache for port 90fbe562-a7de-4ce8-a337-3c3f7be95120 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 985.034939] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.082482] env[61970]: DEBUG oslo_concurrency.lockutils [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.757s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.085427] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.909s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.085699] env[61970]: DEBUG nova.objects.instance [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lazy-loading 'pci_requests' on Instance uuid 7f59000a-94d9-45b6-aa7a-300d95793615 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.101355] env[61970]: INFO nova.scheduler.client.report [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Deleted allocations for instance ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc [ 985.102711] env[61970]: DEBUG oslo_concurrency.lockutils [None req-401c97e4-fe3b-421d-9161-e4fb3bc25919 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "3b1e8446-7862-4532-b2ed-611b63dbf391" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.437s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.431263] env[61970]: DEBUG nova.compute.manager [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 985.589894] env[61970]: DEBUG nova.objects.instance [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lazy-loading 'numa_topology' on Instance uuid 7f59000a-94d9-45b6-aa7a-300d95793615 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.609479] env[61970]: DEBUG oslo_concurrency.lockutils [None req-257aaffa-efa6-4bc3-8611-19d65ba99635 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.431s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.699506] env[61970]: DEBUG nova.network.neutron [req-21133f13-1e98-4ec3-8433-50925a744921 req-73c5976d-c111-4b80-8f7e-fc6e58adabc0 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Added VIF to instance network info cache for port 90fbe562-a7de-4ce8-a337-3c3f7be95120. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3546}} [ 985.699948] env[61970]: DEBUG nova.network.neutron [req-21133f13-1e98-4ec3-8433-50925a744921 req-73c5976d-c111-4b80-8f7e-fc6e58adabc0 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Updating instance_info_cache with network_info: [{"id": "5b506f12-7feb-41e2-8e18-abc9262ac87b", "address": "fa:16:3e:f9:43:05", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b506f12-7f", "ovs_interfaceid": "5b506f12-7feb-41e2-8e18-abc9262ac87b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "90fbe562-a7de-4ce8-a337-3c3f7be95120", "address": "fa:16:3e:91:3a:b6", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90fbe562-a7", "ovs_interfaceid": "90fbe562-a7de-4ce8-a337-3c3f7be95120", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.954094] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.093467] env[61970]: INFO nova.compute.claims [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 986.127639] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "3b1e8446-7862-4532-b2ed-611b63dbf391" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.127639] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "3b1e8446-7862-4532-b2ed-611b63dbf391" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.127816] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "3b1e8446-7862-4532-b2ed-611b63dbf391-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.127945] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "3b1e8446-7862-4532-b2ed-611b63dbf391-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.128165] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "3b1e8446-7862-4532-b2ed-611b63dbf391-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.130169] env[61970]: INFO nova.compute.manager [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Terminating instance [ 986.131980] env[61970]: DEBUG nova.compute.manager [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 986.132193] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 986.133267] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89eee0e-9808-4840-a4a1-be473bf87111 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.141556] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 986.141835] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5fd8a4a0-0410-43ef-afa7-4545fc4d3393 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.148620] env[61970]: DEBUG oslo_vmware.api [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 986.148620] env[61970]: value = "task-1355994" [ 986.148620] env[61970]: _type = "Task" [ 986.148620] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.156706] env[61970]: DEBUG oslo_vmware.api [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355994, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.203962] env[61970]: DEBUG oslo_concurrency.lockutils [req-21133f13-1e98-4ec3-8433-50925a744921 req-73c5976d-c111-4b80-8f7e-fc6e58adabc0 service nova] Releasing lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.204904] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.205308] env[61970]: DEBUG nova.network.neutron [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.660161] env[61970]: DEBUG oslo_vmware.api [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355994, 'name': PowerOffVM_Task, 'duration_secs': 0.284002} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.660585] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 986.660755] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 986.661089] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5f69890c-5146-4c42-9d2b-883c747a67b6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.739650] env[61970]: WARNING nova.network.neutron [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] 8f622fc5-b74f-4582-a6e0-44c90f495750 already exists in list: networks containing: ['8f622fc5-b74f-4582-a6e0-44c90f495750']. ignoring it [ 986.739886] env[61970]: WARNING nova.network.neutron [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] 8f622fc5-b74f-4582-a6e0-44c90f495750 already exists in list: networks containing: ['8f622fc5-b74f-4582-a6e0-44c90f495750']. ignoring it [ 986.740080] env[61970]: WARNING nova.network.neutron [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] 90fbe562-a7de-4ce8-a337-3c3f7be95120 already exists in list: port_ids containing: ['90fbe562-a7de-4ce8-a337-3c3f7be95120']. ignoring it [ 986.748319] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 986.748571] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 986.748772] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Deleting the datastore file [datastore2] 3b1e8446-7862-4532-b2ed-611b63dbf391 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 986.749181] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-909ecf3c-5b46-40db-bc2c-946683f20b35 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.754966] env[61970]: DEBUG oslo_vmware.api [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for the task: (returnval){ [ 986.754966] env[61970]: value = "task-1355996" [ 986.754966] env[61970]: _type = "Task" [ 986.754966] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.762691] env[61970]: DEBUG oslo_vmware.api [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355996, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.963801] env[61970]: DEBUG oslo_concurrency.lockutils [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.964182] env[61970]: DEBUG oslo_concurrency.lockutils [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.119027] env[61970]: DEBUG nova.network.neutron [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Updating instance_info_cache with network_info: [{"id": "5b506f12-7feb-41e2-8e18-abc9262ac87b", "address": "fa:16:3e:f9:43:05", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b506f12-7f", "ovs_interfaceid": "5b506f12-7feb-41e2-8e18-abc9262ac87b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "90fbe562-a7de-4ce8-a337-3c3f7be95120", "address": "fa:16:3e:91:3a:b6", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90fbe562-a7", "ovs_interfaceid": "90fbe562-a7de-4ce8-a337-3c3f7be95120", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.241887] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a28c116-70dc-41ff-ba1a-451e033400e2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.249189] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-864749ce-9282-473d-bf24-f6450abfa959 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.281610] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e439f501-8b67-4e5c-a99d-bc6a0b5dfdef {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.289473] env[61970]: DEBUG oslo_vmware.api [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Task: {'id': task-1355996, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157474} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.291547] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 987.291748] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 987.291930] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 987.292123] env[61970]: INFO nova.compute.manager [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Took 1.16 seconds to destroy the instance on the hypervisor. [ 987.292380] env[61970]: DEBUG oslo.service.loopingcall [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.292719] env[61970]: DEBUG nova.compute.manager [-] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 987.292834] env[61970]: DEBUG nova.network.neutron [-] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 987.295206] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc1413e8-c0a4-4994-95ff-0aedbe456b4a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.309227] env[61970]: DEBUG nova.compute.provider_tree [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.467384] env[61970]: INFO nova.compute.manager [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Detaching volume 148b4432-023f-47fd-b4c0-7e51f31edf89 [ 987.510871] env[61970]: INFO nova.virt.block_device [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Attempting to driver detach volume 148b4432-023f-47fd-b4c0-7e51f31edf89 from mountpoint /dev/sdb [ 987.511428] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Volume detach. Driver type: vmdk {{(pid=61970) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 987.511793] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288827', 'volume_id': '148b4432-023f-47fd-b4c0-7e51f31edf89', 'name': 'volume-148b4432-023f-47fd-b4c0-7e51f31edf89', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa', 'attached_at': '', 'detached_at': '', 'volume_id': '148b4432-023f-47fd-b4c0-7e51f31edf89', 'serial': '148b4432-023f-47fd-b4c0-7e51f31edf89'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 987.517022] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a7ca2a1-b524-4927-9498-7118aa016eb7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.545104] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87504b08-be0c-401e-9e9d-677c38358dc8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.550275] env[61970]: DEBUG nova.compute.manager [req-d30b1499-115d-4bbc-a78e-1fc893381db0 req-c3ee4779-59eb-413b-ae4d-bd705f8110a4 service nova] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Received event network-vif-deleted-a10357cb-17f1-4e77-ae7f-ac439a804c9c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.550553] env[61970]: INFO nova.compute.manager [req-d30b1499-115d-4bbc-a78e-1fc893381db0 req-c3ee4779-59eb-413b-ae4d-bd705f8110a4 service nova] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Neutron deleted interface a10357cb-17f1-4e77-ae7f-ac439a804c9c; detaching it from the instance and deleting it from the info cache [ 987.550767] env[61970]: DEBUG nova.network.neutron [req-d30b1499-115d-4bbc-a78e-1fc893381db0 req-c3ee4779-59eb-413b-ae4d-bd705f8110a4 service nova] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.558113] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c71763e0-1a45-4b9b-ac10-8ca49bcc9834 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.579621] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232a7f88-a7a0-4f8a-9e30-7a1086a39ee1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.594851] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] The volume has not been displaced from its original location: [datastore2] volume-148b4432-023f-47fd-b4c0-7e51f31edf89/volume-148b4432-023f-47fd-b4c0-7e51f31edf89.vmdk. No consolidation needed. {{(pid=61970) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 987.600164] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Reconfiguring VM instance instance-00000048 to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 987.600949] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-639488b5-1b83-4ae9-a639-4162f88b1bf6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.619071] env[61970]: DEBUG oslo_vmware.api [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 987.619071] env[61970]: value = "task-1355997" [ 987.619071] env[61970]: _type = "Task" [ 987.619071] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.622673] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.623290] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.623452] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.624319] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed07c8a-1a99-43db-8579-a4a17d254487 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.631885] env[61970]: DEBUG oslo_vmware.api [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355997, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.643433] env[61970]: DEBUG nova.virt.hardware [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 987.643667] env[61970]: DEBUG nova.virt.hardware [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 987.643830] env[61970]: DEBUG nova.virt.hardware [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.644028] env[61970]: DEBUG nova.virt.hardware [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 987.644183] env[61970]: DEBUG nova.virt.hardware [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.644332] env[61970]: DEBUG nova.virt.hardware [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 987.644544] env[61970]: DEBUG nova.virt.hardware [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 987.644709] env[61970]: DEBUG nova.virt.hardware [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 987.644878] env[61970]: DEBUG nova.virt.hardware [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 987.645054] env[61970]: DEBUG nova.virt.hardware [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 987.645234] env[61970]: DEBUG nova.virt.hardware [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.651439] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Reconfiguring VM to attach interface {{(pid=61970) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 987.651776] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9d6989f-0232-4472-8dd2-1035cea4640d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.668390] env[61970]: DEBUG oslo_vmware.api [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 987.668390] env[61970]: value = "task-1355998" [ 987.668390] env[61970]: _type = "Task" [ 987.668390] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.676249] env[61970]: DEBUG oslo_vmware.api [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355998, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.813302] env[61970]: DEBUG nova.scheduler.client.report [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 988.006528] env[61970]: DEBUG nova.network.neutron [-] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.054055] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-81d5db63-d580-4be7-81d9-7c45db85cfae {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.063155] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d57fecf-255b-444d-bc0e-87134d12fdcf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.090789] env[61970]: DEBUG nova.compute.manager [req-d30b1499-115d-4bbc-a78e-1fc893381db0 req-c3ee4779-59eb-413b-ae4d-bd705f8110a4 service nova] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Detach interface failed, port_id=a10357cb-17f1-4e77-ae7f-ac439a804c9c, reason: Instance 3b1e8446-7862-4532-b2ed-611b63dbf391 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 988.128621] env[61970]: DEBUG oslo_vmware.api [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355997, 'name': ReconfigVM_Task, 'duration_secs': 0.248541} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.128951] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Reconfigured VM instance instance-00000048 to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 988.134024] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c9207ca-5495-4a18-ac17-cc406fae6b4d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.150021] env[61970]: DEBUG oslo_vmware.api [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 988.150021] env[61970]: value = "task-1355999" [ 988.150021] env[61970]: _type = "Task" [ 988.150021] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.157730] env[61970]: DEBUG oslo_vmware.api [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355999, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.177652] env[61970]: DEBUG oslo_vmware.api [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355998, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.319140] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.233s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.321104] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.414s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.322685] env[61970]: INFO nova.compute.claims [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 988.363878] env[61970]: INFO nova.network.neutron [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Updating port 5c475b00-f4ed-4e09-916c-6f3ca7845784 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 988.509353] env[61970]: INFO nova.compute.manager [-] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Took 1.22 seconds to deallocate network for instance. [ 988.661644] env[61970]: DEBUG oslo_vmware.api [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355999, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.678774] env[61970]: DEBUG oslo_vmware.api [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355998, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.015732] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.160412] env[61970]: DEBUG oslo_vmware.api [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355999, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.178090] env[61970]: DEBUG oslo_vmware.api [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1355998, 'name': ReconfigVM_Task, 'duration_secs': 1.162937} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.178641] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.178877] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Reconfigured VM to attach interface {{(pid=61970) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 989.476848] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b48246-3282-44b3-b194-23373b19cdfd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.484508] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ceb8de6-e532-4bd3-8c82-76f66a62ec65 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.514123] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e6d718-1ff5-4092-be4f-704add51b3eb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.522101] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26378ed5-76ed-4461-9324-60a1c9e6aa59 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.535470] env[61970]: DEBUG nova.compute.provider_tree [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.661066] env[61970]: DEBUG oslo_vmware.api [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1355999, 'name': ReconfigVM_Task, 'duration_secs': 1.158713} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.661384] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288827', 'volume_id': '148b4432-023f-47fd-b4c0-7e51f31edf89', 'name': 'volume-148b4432-023f-47fd-b4c0-7e51f31edf89', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa', 'attached_at': '', 'detached_at': '', 'volume_id': '148b4432-023f-47fd-b4c0-7e51f31edf89', 'serial': '148b4432-023f-47fd-b4c0-7e51f31edf89'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 989.683532] env[61970]: DEBUG oslo_concurrency.lockutils [None req-eafff75a-684f-4b62-a32a-544db502f65d tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "interface-c4ec1797-d939-4c24-9314-ce9c74eb5aa1-90fbe562-a7de-4ce8-a337-3c3f7be95120" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.893s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.793287] env[61970]: DEBUG nova.compute.manager [req-2d677057-0b8c-4ce9-bea7-8af22a380b5e req-d3c359bb-b073-476d-9697-ad7c8477ef80 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Received event network-vif-plugged-5c475b00-f4ed-4e09-916c-6f3ca7845784 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.793559] env[61970]: DEBUG oslo_concurrency.lockutils [req-2d677057-0b8c-4ce9-bea7-8af22a380b5e req-d3c359bb-b073-476d-9697-ad7c8477ef80 service nova] Acquiring lock "7f59000a-94d9-45b6-aa7a-300d95793615-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.793774] env[61970]: DEBUG oslo_concurrency.lockutils [req-2d677057-0b8c-4ce9-bea7-8af22a380b5e req-d3c359bb-b073-476d-9697-ad7c8477ef80 service nova] Lock "7f59000a-94d9-45b6-aa7a-300d95793615-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.793951] env[61970]: DEBUG oslo_concurrency.lockutils [req-2d677057-0b8c-4ce9-bea7-8af22a380b5e req-d3c359bb-b073-476d-9697-ad7c8477ef80 service nova] Lock "7f59000a-94d9-45b6-aa7a-300d95793615-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.794140] env[61970]: DEBUG nova.compute.manager [req-2d677057-0b8c-4ce9-bea7-8af22a380b5e req-d3c359bb-b073-476d-9697-ad7c8477ef80 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] No waiting events found dispatching network-vif-plugged-5c475b00-f4ed-4e09-916c-6f3ca7845784 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 989.794313] env[61970]: WARNING nova.compute.manager [req-2d677057-0b8c-4ce9-bea7-8af22a380b5e req-d3c359bb-b073-476d-9697-ad7c8477ef80 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Received unexpected event network-vif-plugged-5c475b00-f4ed-4e09-916c-6f3ca7845784 for instance with vm_state shelved_offloaded and task_state spawning. [ 989.884464] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.884714] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquired lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.884844] env[61970]: DEBUG nova.network.neutron [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 990.038405] env[61970]: DEBUG nova.scheduler.client.report [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.204887] env[61970]: DEBUG nova.objects.instance [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lazy-loading 'flavor' on Instance uuid 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.544588] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.222s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.544588] env[61970]: DEBUG nova.compute.manager [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 990.546671] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.593s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.548060] env[61970]: INFO nova.compute.claims [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 990.619873] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3bc9ef10-9ed9-499b-8dcc-8da188fe68fa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.620153] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3bc9ef10-9ed9-499b-8dcc-8da188fe68fa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.663157] env[61970]: DEBUG nova.network.neutron [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Updating instance_info_cache with network_info: [{"id": "5c475b00-f4ed-4e09-916c-6f3ca7845784", "address": "fa:16:3e:27:f3:40", "network": {"id": "70081995-3d6c-466e-b811-1c8e532e552f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-641102750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1d4724cbb0841859369af01f2c03db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aaf1b231-3660-4453-b4f3-44d825b9a5dd", "external-id": "nsx-vlan-transportzone-6", "segmentation_id": 6, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c475b00-f4", "ovs_interfaceid": "5c475b00-f4ed-4e09-916c-6f3ca7845784", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.042064] env[61970]: DEBUG oslo_concurrency.lockutils [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "interface-c4ec1797-d939-4c24-9314-ce9c74eb5aa1-90fbe562-a7de-4ce8-a337-3c3f7be95120" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.042493] env[61970]: DEBUG oslo_concurrency.lockutils [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "interface-c4ec1797-d939-4c24-9314-ce9c74eb5aa1-90fbe562-a7de-4ce8-a337-3c3f7be95120" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.051926] env[61970]: DEBUG nova.compute.utils [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 991.055904] env[61970]: DEBUG nova.compute.manager [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 991.055904] env[61970]: DEBUG nova.network.neutron [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 991.095171] env[61970]: DEBUG nova.policy [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64082edf5f2e4621b5f3586e1792e42a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9235eeeb6fa4ccf959ffbea456f3694', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 991.124008] env[61970]: INFO nova.compute.manager [None req-3bc9ef10-9ed9-499b-8dcc-8da188fe68fa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Detaching volume 3916ae0c-f673-4a6a-bd2a-9134a92569c0 [ 991.158679] env[61970]: INFO nova.virt.block_device [None req-3bc9ef10-9ed9-499b-8dcc-8da188fe68fa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Attempting to driver detach volume 3916ae0c-f673-4a6a-bd2a-9134a92569c0 from mountpoint /dev/sdb [ 991.158985] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bc9ef10-9ed9-499b-8dcc-8da188fe68fa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Volume detach. Driver type: vmdk {{(pid=61970) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 991.159199] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bc9ef10-9ed9-499b-8dcc-8da188fe68fa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288840', 'volume_id': '3916ae0c-f673-4a6a-bd2a-9134a92569c0', 'name': 'volume-3916ae0c-f673-4a6a-bd2a-9134a92569c0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd0ad9689-bdb7-4d68-b19a-3f805ab4612e', 'attached_at': '', 'detached_at': '', 'volume_id': '3916ae0c-f673-4a6a-bd2a-9134a92569c0', 'serial': '3916ae0c-f673-4a6a-bd2a-9134a92569c0'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 991.160096] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2741aa-e7d0-4638-9998-7febe2410051 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.182914] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Releasing lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.185754] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b269d3bd-9d71-4de0-bb64-77b863017f18 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.192554] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf3afe3f-6b0f-4d3e-b873-7720a641c20b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.212797] env[61970]: DEBUG oslo_concurrency.lockutils [None req-68fdaad7-e0f1-4122-9371-193f39abeb28 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.249s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.216304] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f681ab4c-0a72-4b79-b35a-e285bffdea60 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.232718] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bc9ef10-9ed9-499b-8dcc-8da188fe68fa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] The volume has not been displaced from its original location: [datastore1] volume-3916ae0c-f673-4a6a-bd2a-9134a92569c0/volume-3916ae0c-f673-4a6a-bd2a-9134a92569c0.vmdk. No consolidation needed. {{(pid=61970) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 991.237962] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bc9ef10-9ed9-499b-8dcc-8da188fe68fa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Reconfiguring VM instance instance-0000004e to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 991.240490] env[61970]: DEBUG nova.virt.hardware [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='6106c698033312a540459499dc7ea060',container_format='bare',created_at=2024-10-24T13:45:08Z,direct_url=,disk_format='vmdk',id=8ff34e61-3203-40b2-805a-fd5355a36d65,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1613976763-shelved',owner='e1d4724cbb0841859369af01f2c03db1',properties=ImageMetaProps,protected=,size=31670784,status='active',tags=,updated_at=2024-10-24T13:45:21Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 991.240748] env[61970]: DEBUG nova.virt.hardware [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 991.240973] env[61970]: DEBUG nova.virt.hardware [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 991.241197] env[61970]: DEBUG nova.virt.hardware [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 991.241286] env[61970]: DEBUG nova.virt.hardware [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 991.241491] env[61970]: DEBUG nova.virt.hardware [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 991.241669] env[61970]: DEBUG nova.virt.hardware [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 991.241864] env[61970]: DEBUG nova.virt.hardware [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 991.242068] env[61970]: DEBUG nova.virt.hardware [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 991.242242] env[61970]: DEBUG nova.virt.hardware [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 991.242425] env[61970]: DEBUG nova.virt.hardware [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 991.243461] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec0bffbf-c449-40ff-bbf6-515f559e74b6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.257424] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b0de0e5-6c9b-4e76-95f3-154bc9926abb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.267489] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-636aa979-e68c-4e79-8c26-1b943a2382d0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.271860] env[61970]: DEBUG oslo_vmware.api [None req-3bc9ef10-9ed9-499b-8dcc-8da188fe68fa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 991.271860] env[61970]: value = "task-1356000" [ 991.271860] env[61970]: _type = "Task" [ 991.271860] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.285753] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:f3:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aaf1b231-3660-4453-b4f3-44d825b9a5dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5c475b00-f4ed-4e09-916c-6f3ca7845784', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 991.293591] env[61970]: DEBUG oslo.service.loopingcall [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.294477] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 991.294610] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b8acd048-4aa3-41eb-a3e2-79fe43748951 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.312491] env[61970]: DEBUG oslo_vmware.api [None req-3bc9ef10-9ed9-499b-8dcc-8da188fe68fa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356000, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.317341] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 991.317341] env[61970]: value = "task-1356001" [ 991.317341] env[61970]: _type = "Task" [ 991.317341] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.324816] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356001, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.370603] env[61970]: DEBUG oslo_concurrency.lockutils [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.370603] env[61970]: DEBUG oslo_concurrency.lockutils [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.370698] env[61970]: DEBUG oslo_concurrency.lockutils [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.373366] env[61970]: DEBUG oslo_concurrency.lockutils [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.373366] env[61970]: DEBUG oslo_concurrency.lockutils [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.373446] env[61970]: INFO nova.compute.manager [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Terminating instance [ 991.377174] env[61970]: DEBUG nova.compute.manager [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 991.377431] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 991.378244] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d713e285-f577-45a4-9e80-1ab2e78a1635 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.386090] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 991.386489] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8185950-4916-495c-a16c-14c7b83e40ff {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.393097] env[61970]: DEBUG oslo_vmware.api [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 991.393097] env[61970]: value = "task-1356002" [ 991.393097] env[61970]: _type = "Task" [ 991.393097] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.402642] env[61970]: DEBUG oslo_vmware.api [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356002, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.424086] env[61970]: DEBUG nova.network.neutron [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Successfully created port: f7943b8e-3fab-4cfc-b59d-5a7a01efb9ba {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 991.545063] env[61970]: DEBUG oslo_concurrency.lockutils [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.545333] env[61970]: DEBUG oslo_concurrency.lockutils [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.546178] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59fd66b1-4a10-4283-a45a-570ed0a460ce {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.568457] env[61970]: DEBUG nova.compute.manager [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 991.572646] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59736eb7-868c-4879-b1c5-c572996404d9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.600529] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Reconfiguring VM to detach interface {{(pid=61970) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 991.603956] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-edafb0f6-1803-4829-82e8-f18f1b223fcf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.625478] env[61970]: DEBUG oslo_vmware.api [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 991.625478] env[61970]: value = "task-1356003" [ 991.625478] env[61970]: _type = "Task" [ 991.625478] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.636216] env[61970]: DEBUG oslo_vmware.api [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356003, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.753438] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f263f4e3-ea27-4dbc-b43b-e0f35bcf9b1c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.760880] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbec6c27-dcb4-4697-8d78-d7e90ac1b14c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.794094] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11bf48a2-51b5-4fbb-aaaf-a6e5ce9fd7d1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.801889] env[61970]: DEBUG oslo_vmware.api [None req-3bc9ef10-9ed9-499b-8dcc-8da188fe68fa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356000, 'name': ReconfigVM_Task, 'duration_secs': 0.328516} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.804106] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bc9ef10-9ed9-499b-8dcc-8da188fe68fa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Reconfigured VM instance instance-0000004e to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 991.808901] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec2f3537-13b2-48f2-bd90-d39421586578 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.820098] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea85b17-b94d-4383-bccd-0f948ee65eb7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.826191] env[61970]: DEBUG nova.compute.manager [req-04746ac5-4507-4c78-9c51-d632356e2c98 req-3ac89712-3cd5-44b9-bf76-7dec3dc44c7e service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Received event network-changed-5c475b00-f4ed-4e09-916c-6f3ca7845784 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.826339] env[61970]: DEBUG nova.compute.manager [req-04746ac5-4507-4c78-9c51-d632356e2c98 req-3ac89712-3cd5-44b9-bf76-7dec3dc44c7e service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Refreshing instance network info cache due to event network-changed-5c475b00-f4ed-4e09-916c-6f3ca7845784. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 991.826560] env[61970]: DEBUG oslo_concurrency.lockutils [req-04746ac5-4507-4c78-9c51-d632356e2c98 req-3ac89712-3cd5-44b9-bf76-7dec3dc44c7e service nova] Acquiring lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.826705] env[61970]: DEBUG oslo_concurrency.lockutils [req-04746ac5-4507-4c78-9c51-d632356e2c98 req-3ac89712-3cd5-44b9-bf76-7dec3dc44c7e service nova] Acquired lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.826867] env[61970]: DEBUG nova.network.neutron [req-04746ac5-4507-4c78-9c51-d632356e2c98 req-3ac89712-3cd5-44b9-bf76-7dec3dc44c7e service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Refreshing network info cache for port 5c475b00-f4ed-4e09-916c-6f3ca7845784 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 991.846391] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356001, 'name': CreateVM_Task, 'duration_secs': 0.436039} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.847125] env[61970]: DEBUG nova.compute.provider_tree [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.849386] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 991.849942] env[61970]: DEBUG oslo_vmware.api [None req-3bc9ef10-9ed9-499b-8dcc-8da188fe68fa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 991.849942] env[61970]: value = "task-1356004" [ 991.849942] env[61970]: _type = "Task" [ 991.849942] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.850749] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8ff34e61-3203-40b2-805a-fd5355a36d65" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.850908] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8ff34e61-3203-40b2-805a-fd5355a36d65" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.851328] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8ff34e61-3203-40b2-805a-fd5355a36d65" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 991.851632] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f22b077-cbe0-460b-9477-d27847fcf32b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.860708] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 991.860708] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52fced2e-fd99-30a9-2b40-82d32ed6b620" [ 991.860708] env[61970]: _type = "Task" [ 991.860708] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.864189] env[61970]: DEBUG oslo_vmware.api [None req-3bc9ef10-9ed9-499b-8dcc-8da188fe68fa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356004, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.871774] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52fced2e-fd99-30a9-2b40-82d32ed6b620, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.901395] env[61970]: DEBUG oslo_vmware.api [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356002, 'name': PowerOffVM_Task, 'duration_secs': 0.21335} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.901650] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 991.901918] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 991.902200] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6ca9f85d-9cfe-4ae2-a81a-e3be6f4ef345 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.971387] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 991.972042] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 991.972042] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Deleting the datastore file [datastore2] 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 991.972448] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14cdb808-c05a-41b7-817f-b4b23a3bf594 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.980232] env[61970]: DEBUG oslo_vmware.api [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 991.980232] env[61970]: value = "task-1356006" [ 991.980232] env[61970]: _type = "Task" [ 991.980232] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.988202] env[61970]: DEBUG oslo_vmware.api [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356006, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.135481] env[61970]: DEBUG oslo_vmware.api [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356003, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.352405] env[61970]: DEBUG nova.scheduler.client.report [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.365763] env[61970]: DEBUG oslo_vmware.api [None req-3bc9ef10-9ed9-499b-8dcc-8da188fe68fa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356004, 'name': ReconfigVM_Task, 'duration_secs': 0.144626} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.369094] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bc9ef10-9ed9-499b-8dcc-8da188fe68fa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288840', 'volume_id': '3916ae0c-f673-4a6a-bd2a-9134a92569c0', 'name': 'volume-3916ae0c-f673-4a6a-bd2a-9134a92569c0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd0ad9689-bdb7-4d68-b19a-3f805ab4612e', 'attached_at': '', 'detached_at': '', 'volume_id': '3916ae0c-f673-4a6a-bd2a-9134a92569c0', 'serial': '3916ae0c-f673-4a6a-bd2a-9134a92569c0'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 992.378012] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8ff34e61-3203-40b2-805a-fd5355a36d65" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.378291] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Processing image 8ff34e61-3203-40b2-805a-fd5355a36d65 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 992.378525] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8ff34e61-3203-40b2-805a-fd5355a36d65/8ff34e61-3203-40b2-805a-fd5355a36d65.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.378676] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8ff34e61-3203-40b2-805a-fd5355a36d65/8ff34e61-3203-40b2-805a-fd5355a36d65.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.378856] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 992.379128] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b243c516-81ae-4d59-afd9-cf2588212da2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.389682] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 992.389889] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 992.390647] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f26895cf-af7a-42a1-83fa-d448f34b0654 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.396565] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 992.396565] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ef3840-0829-772c-d240-1999ad59a505" [ 992.396565] env[61970]: _type = "Task" [ 992.396565] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.410181] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ef3840-0829-772c-d240-1999ad59a505, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.490627] env[61970]: DEBUG oslo_vmware.api [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356006, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173031} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.493160] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.493399] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 992.493568] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 992.494206] env[61970]: INFO nova.compute.manager [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Took 1.12 seconds to destroy the instance on the hypervisor. [ 992.494206] env[61970]: DEBUG oslo.service.loopingcall [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.494345] env[61970]: DEBUG nova.compute.manager [-] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 992.494385] env[61970]: DEBUG nova.network.neutron [-] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 992.527927] env[61970]: DEBUG oslo_vmware.rw_handles [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525d1550-fb59-aae7-beee-d43377562f35/disk-0.vmdk. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 992.528967] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422c9140-9933-40b7-9de4-30c71bdef6a5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.535160] env[61970]: DEBUG oslo_vmware.rw_handles [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525d1550-fb59-aae7-beee-d43377562f35/disk-0.vmdk is in state: ready. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 992.535331] env[61970]: ERROR oslo_vmware.rw_handles [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525d1550-fb59-aae7-beee-d43377562f35/disk-0.vmdk due to incomplete transfer. [ 992.535559] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1d608dd5-90ab-4f12-825c-5268241b1dd0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.544944] env[61970]: DEBUG oslo_vmware.rw_handles [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525d1550-fb59-aae7-beee-d43377562f35/disk-0.vmdk. {{(pid=61970) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 992.545171] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Uploaded image de2205ef-3756-4fc4-896c-06f28059131f to the Glance image server {{(pid=61970) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 992.547396] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Destroying the VM {{(pid=61970) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 992.547650] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-455bfe3d-a27b-4f2e-94b3-a2713c9af749 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.557593] env[61970]: DEBUG oslo_vmware.api [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 992.557593] env[61970]: value = "task-1356007" [ 992.557593] env[61970]: _type = "Task" [ 992.557593] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.565239] env[61970]: DEBUG oslo_vmware.api [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356007, 'name': Destroy_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.582011] env[61970]: DEBUG nova.compute.manager [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 992.603080] env[61970]: DEBUG nova.virt.hardware [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 992.603080] env[61970]: DEBUG nova.virt.hardware [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 992.603285] env[61970]: DEBUG nova.virt.hardware [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 992.603486] env[61970]: DEBUG nova.virt.hardware [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 992.603668] env[61970]: DEBUG nova.virt.hardware [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 992.603844] env[61970]: DEBUG nova.virt.hardware [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 992.604329] env[61970]: DEBUG nova.virt.hardware [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 992.604585] env[61970]: DEBUG nova.virt.hardware [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 992.604774] env[61970]: DEBUG nova.virt.hardware [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 992.604945] env[61970]: DEBUG nova.virt.hardware [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 992.605139] env[61970]: DEBUG nova.virt.hardware [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 992.606440] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa19e06-2e66-4162-85c7-716657d41157 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.615563] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db5d2bf-e077-4a86-9280-9a13b9969f66 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.643058] env[61970]: DEBUG oslo_vmware.api [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356003, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.670567] env[61970]: DEBUG nova.network.neutron [req-04746ac5-4507-4c78-9c51-d632356e2c98 req-3ac89712-3cd5-44b9-bf76-7dec3dc44c7e service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Updated VIF entry in instance network info cache for port 5c475b00-f4ed-4e09-916c-6f3ca7845784. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 992.671099] env[61970]: DEBUG nova.network.neutron [req-04746ac5-4507-4c78-9c51-d632356e2c98 req-3ac89712-3cd5-44b9-bf76-7dec3dc44c7e service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Updating instance_info_cache with network_info: [{"id": "5c475b00-f4ed-4e09-916c-6f3ca7845784", "address": "fa:16:3e:27:f3:40", "network": {"id": "70081995-3d6c-466e-b811-1c8e532e552f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-641102750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1d4724cbb0841859369af01f2c03db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aaf1b231-3660-4453-b4f3-44d825b9a5dd", "external-id": "nsx-vlan-transportzone-6", "segmentation_id": 6, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c475b00-f4", "ovs_interfaceid": "5c475b00-f4ed-4e09-916c-6f3ca7845784", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.860529] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.314s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.861084] env[61970]: DEBUG nova.compute.manager [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 992.863701] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.848s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.867022] env[61970]: DEBUG nova.objects.instance [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lazy-loading 'resources' on Instance uuid 3b1e8446-7862-4532-b2ed-611b63dbf391 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.907615] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Preparing fetch location {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 992.907901] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Fetch image to [datastore2] OSTACK_IMG_c0a6eb66-edfc-4d4a-9b6a-ccd264c124b8/OSTACK_IMG_c0a6eb66-edfc-4d4a-9b6a-ccd264c124b8.vmdk {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 992.908119] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Downloading stream optimized image 8ff34e61-3203-40b2-805a-fd5355a36d65 to [datastore2] OSTACK_IMG_c0a6eb66-edfc-4d4a-9b6a-ccd264c124b8/OSTACK_IMG_c0a6eb66-edfc-4d4a-9b6a-ccd264c124b8.vmdk on the data store datastore2 as vApp {{(pid=61970) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 992.908386] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Downloading image file data 8ff34e61-3203-40b2-805a-fd5355a36d65 to the ESX as VM named 'OSTACK_IMG_c0a6eb66-edfc-4d4a-9b6a-ccd264c124b8' {{(pid=61970) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 992.923750] env[61970]: DEBUG nova.objects.instance [None req-3bc9ef10-9ed9-499b-8dcc-8da188fe68fa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lazy-loading 'flavor' on Instance uuid d0ad9689-bdb7-4d68-b19a-3f805ab4612e {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.987664] env[61970]: DEBUG oslo_vmware.rw_handles [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 992.987664] env[61970]: value = "resgroup-9" [ 992.987664] env[61970]: _type = "ResourcePool" [ 992.987664] env[61970]: }. {{(pid=61970) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 992.987953] env[61970]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-47f90c09-e088-4cc9-8cd1-76e5af5995e8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.012252] env[61970]: DEBUG oslo_vmware.rw_handles [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lease: (returnval){ [ 993.012252] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529ee39b-aaae-f730-9303-b5cdbb96fa41" [ 993.012252] env[61970]: _type = "HttpNfcLease" [ 993.012252] env[61970]: } obtained for vApp import into resource pool (val){ [ 993.012252] env[61970]: value = "resgroup-9" [ 993.012252] env[61970]: _type = "ResourcePool" [ 993.012252] env[61970]: }. {{(pid=61970) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 993.012557] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the lease: (returnval){ [ 993.012557] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529ee39b-aaae-f730-9303-b5cdbb96fa41" [ 993.012557] env[61970]: _type = "HttpNfcLease" [ 993.012557] env[61970]: } to be ready. {{(pid=61970) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 993.019776] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 993.019776] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529ee39b-aaae-f730-9303-b5cdbb96fa41" [ 993.019776] env[61970]: _type = "HttpNfcLease" [ 993.019776] env[61970]: } is initializing. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 993.066639] env[61970]: DEBUG oslo_vmware.api [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356007, 'name': Destroy_Task, 'duration_secs': 0.301326} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.066919] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Destroyed the VM [ 993.067181] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Deleting Snapshot of the VM instance {{(pid=61970) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 993.067515] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9b01df6c-6591-4e26-87a0-8a10ef68bc2c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.073634] env[61970]: DEBUG oslo_vmware.api [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 993.073634] env[61970]: value = "task-1356009" [ 993.073634] env[61970]: _type = "Task" [ 993.073634] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.081285] env[61970]: DEBUG oslo_vmware.api [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356009, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.118421] env[61970]: DEBUG nova.network.neutron [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Successfully updated port: f7943b8e-3fab-4cfc-b59d-5a7a01efb9ba {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 993.145094] env[61970]: DEBUG oslo_vmware.api [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356003, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.178649] env[61970]: DEBUG oslo_concurrency.lockutils [req-04746ac5-4507-4c78-9c51-d632356e2c98 req-3ac89712-3cd5-44b9-bf76-7dec3dc44c7e service nova] Releasing lock "refresh_cache-7f59000a-94d9-45b6-aa7a-300d95793615" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.367206] env[61970]: DEBUG nova.compute.utils [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 993.371798] env[61970]: DEBUG nova.compute.manager [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 993.372552] env[61970]: DEBUG nova.network.neutron [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 993.412372] env[61970]: DEBUG nova.policy [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd04344afddb14418b5018b69fbd7694e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c6a4997482a4834a30c79c4cbab2cc1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 993.520388] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 993.520388] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529ee39b-aaae-f730-9303-b5cdbb96fa41" [ 993.520388] env[61970]: _type = "HttpNfcLease" [ 993.520388] env[61970]: } is initializing. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 993.534181] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a21cb383-eb70-4b8c-89a3-a3a5f1244297 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.541730] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e5fa02a-8575-4b61-92e2-84700fb42e7b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.571442] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e70b787b-80f8-449e-a19b-109a38e4d6fe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.584705] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d898f4e0-3c10-4715-900e-48082ef319a8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.588450] env[61970]: DEBUG oslo_vmware.api [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356009, 'name': RemoveSnapshot_Task} progress is 17%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.597932] env[61970]: DEBUG nova.compute.provider_tree [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.621348] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "refresh_cache-00e3287d-24fb-4ea9-b212-54215828d5df" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.621521] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "refresh_cache-00e3287d-24fb-4ea9-b212-54215828d5df" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.621658] env[61970]: DEBUG nova.network.neutron [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 993.646861] env[61970]: DEBUG oslo_vmware.api [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356003, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.680340] env[61970]: DEBUG nova.network.neutron [-] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.686247] env[61970]: DEBUG nova.network.neutron [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Successfully created port: 62d22328-b863-41dc-816f-1ce3f7a53eed {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 993.799420] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4f2e7e03-53c0-490a-b358-6cfcc0485979 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.869107] env[61970]: DEBUG nova.compute.manager [req-3998eab5-884c-4140-87d3-418f3fdbaf01 req-b9188a7d-536e-40a7-878f-ce7aab9c059d service nova] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Received event network-vif-plugged-f7943b8e-3fab-4cfc-b59d-5a7a01efb9ba {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.869896] env[61970]: DEBUG oslo_concurrency.lockutils [req-3998eab5-884c-4140-87d3-418f3fdbaf01 req-b9188a7d-536e-40a7-878f-ce7aab9c059d service nova] Acquiring lock "00e3287d-24fb-4ea9-b212-54215828d5df-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.870372] env[61970]: DEBUG oslo_concurrency.lockutils [req-3998eab5-884c-4140-87d3-418f3fdbaf01 req-b9188a7d-536e-40a7-878f-ce7aab9c059d service nova] Lock "00e3287d-24fb-4ea9-b212-54215828d5df-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.873109] env[61970]: DEBUG oslo_concurrency.lockutils [req-3998eab5-884c-4140-87d3-418f3fdbaf01 req-b9188a7d-536e-40a7-878f-ce7aab9c059d service nova] Lock "00e3287d-24fb-4ea9-b212-54215828d5df-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.873109] env[61970]: DEBUG nova.compute.manager [req-3998eab5-884c-4140-87d3-418f3fdbaf01 req-b9188a7d-536e-40a7-878f-ce7aab9c059d service nova] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] No waiting events found dispatching network-vif-plugged-f7943b8e-3fab-4cfc-b59d-5a7a01efb9ba {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 993.873109] env[61970]: WARNING nova.compute.manager [req-3998eab5-884c-4140-87d3-418f3fdbaf01 req-b9188a7d-536e-40a7-878f-ce7aab9c059d service nova] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Received unexpected event network-vif-plugged-f7943b8e-3fab-4cfc-b59d-5a7a01efb9ba for instance with vm_state building and task_state spawning. [ 993.873109] env[61970]: DEBUG nova.compute.manager [req-3998eab5-884c-4140-87d3-418f3fdbaf01 req-b9188a7d-536e-40a7-878f-ce7aab9c059d service nova] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Received event network-changed-f7943b8e-3fab-4cfc-b59d-5a7a01efb9ba {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.873109] env[61970]: DEBUG nova.compute.manager [req-3998eab5-884c-4140-87d3-418f3fdbaf01 req-b9188a7d-536e-40a7-878f-ce7aab9c059d service nova] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Refreshing instance network info cache due to event network-changed-f7943b8e-3fab-4cfc-b59d-5a7a01efb9ba. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 993.873109] env[61970]: DEBUG oslo_concurrency.lockutils [req-3998eab5-884c-4140-87d3-418f3fdbaf01 req-b9188a7d-536e-40a7-878f-ce7aab9c059d service nova] Acquiring lock "refresh_cache-00e3287d-24fb-4ea9-b212-54215828d5df" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.873109] env[61970]: DEBUG nova.compute.manager [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 993.934014] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3bc9ef10-9ed9-499b-8dcc-8da188fe68fa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.312s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.934014] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4f2e7e03-53c0-490a-b358-6cfcc0485979 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.135s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.934511] env[61970]: DEBUG nova.compute.manager [None req-4f2e7e03-53c0-490a-b358-6cfcc0485979 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 993.936080] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c21c1c7d-e12b-44a2-a8d2-b4cfa90fa98f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.948270] env[61970]: DEBUG nova.compute.manager [None req-4f2e7e03-53c0-490a-b358-6cfcc0485979 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61970) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 993.949877] env[61970]: DEBUG nova.objects.instance [None req-4f2e7e03-53c0-490a-b358-6cfcc0485979 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lazy-loading 'flavor' on Instance uuid d0ad9689-bdb7-4d68-b19a-3f805ab4612e {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 994.022771] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 994.022771] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529ee39b-aaae-f730-9303-b5cdbb96fa41" [ 994.022771] env[61970]: _type = "HttpNfcLease" [ 994.022771] env[61970]: } is initializing. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 994.085972] env[61970]: DEBUG oslo_vmware.api [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356009, 'name': RemoveSnapshot_Task, 'duration_secs': 0.895768} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.085972] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Deleted Snapshot of the VM instance {{(pid=61970) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 994.085972] env[61970]: INFO nova.compute.manager [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Took 15.23 seconds to snapshot the instance on the hypervisor. [ 994.101785] env[61970]: DEBUG nova.scheduler.client.report [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 994.146161] env[61970]: DEBUG oslo_vmware.api [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356003, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.153999] env[61970]: DEBUG nova.network.neutron [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 994.184263] env[61970]: INFO nova.compute.manager [-] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Took 1.69 seconds to deallocate network for instance. [ 994.300909] env[61970]: DEBUG nova.network.neutron [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Updating instance_info_cache with network_info: [{"id": "f7943b8e-3fab-4cfc-b59d-5a7a01efb9ba", "address": "fa:16:3e:70:b8:6e", "network": {"id": "e02bfa6c-8f0d-4413-8dcf-e644c07792d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1410864732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9235eeeb6fa4ccf959ffbea456f3694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7943b8e-3f", "ovs_interfaceid": "f7943b8e-3fab-4cfc-b59d-5a7a01efb9ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.456397] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f2e7e03-53c0-490a-b358-6cfcc0485979 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 994.456397] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-74279a4b-e43a-472d-9cda-bdf306e7ce92 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.464125] env[61970]: DEBUG oslo_vmware.api [None req-4f2e7e03-53c0-490a-b358-6cfcc0485979 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 994.464125] env[61970]: value = "task-1356010" [ 994.464125] env[61970]: _type = "Task" [ 994.464125] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.473051] env[61970]: DEBUG oslo_vmware.api [None req-4f2e7e03-53c0-490a-b358-6cfcc0485979 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356010, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.522375] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 994.522375] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529ee39b-aaae-f730-9303-b5cdbb96fa41" [ 994.522375] env[61970]: _type = "HttpNfcLease" [ 994.522375] env[61970]: } is initializing. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 994.606929] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.743s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.631398] env[61970]: INFO nova.scheduler.client.report [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Deleted allocations for instance 3b1e8446-7862-4532-b2ed-611b63dbf391 [ 994.634753] env[61970]: DEBUG nova.compute.manager [None req-9af09210-ed2e-4d09-ba7f-0f1385604a2e tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Found 2 images (rotation: 2) {{(pid=61970) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 994.652531] env[61970]: DEBUG oslo_vmware.api [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356003, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.692871] env[61970]: DEBUG oslo_concurrency.lockutils [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.693241] env[61970]: DEBUG oslo_concurrency.lockutils [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.693508] env[61970]: DEBUG nova.objects.instance [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lazy-loading 'resources' on Instance uuid 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 994.803687] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "refresh_cache-00e3287d-24fb-4ea9-b212-54215828d5df" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.803962] env[61970]: DEBUG nova.compute.manager [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Instance network_info: |[{"id": "f7943b8e-3fab-4cfc-b59d-5a7a01efb9ba", "address": "fa:16:3e:70:b8:6e", "network": {"id": "e02bfa6c-8f0d-4413-8dcf-e644c07792d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1410864732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9235eeeb6fa4ccf959ffbea456f3694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7943b8e-3f", "ovs_interfaceid": "f7943b8e-3fab-4cfc-b59d-5a7a01efb9ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 994.804307] env[61970]: DEBUG oslo_concurrency.lockutils [req-3998eab5-884c-4140-87d3-418f3fdbaf01 req-b9188a7d-536e-40a7-878f-ce7aab9c059d service nova] Acquired lock "refresh_cache-00e3287d-24fb-4ea9-b212-54215828d5df" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.804494] env[61970]: DEBUG nova.network.neutron [req-3998eab5-884c-4140-87d3-418f3fdbaf01 req-b9188a7d-536e-40a7-878f-ce7aab9c059d service nova] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Refreshing network info cache for port f7943b8e-3fab-4cfc-b59d-5a7a01efb9ba {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 994.807645] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:b8:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '510d3c47-3615-43d5-aa5d-a279fd915e71', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f7943b8e-3fab-4cfc-b59d-5a7a01efb9ba', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 994.815693] env[61970]: DEBUG oslo.service.loopingcall [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.816219] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 994.817033] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7791cc32-1d91-4f96-9da0-0272993e30ac {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.837233] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 994.837233] env[61970]: value = "task-1356011" [ 994.837233] env[61970]: _type = "Task" [ 994.837233] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.845357] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356011, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.883276] env[61970]: DEBUG nova.compute.manager [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 994.909632] env[61970]: DEBUG nova.virt.hardware [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 994.909912] env[61970]: DEBUG nova.virt.hardware [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 994.910089] env[61970]: DEBUG nova.virt.hardware [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 994.910395] env[61970]: DEBUG nova.virt.hardware [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 994.910568] env[61970]: DEBUG nova.virt.hardware [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 994.910721] env[61970]: DEBUG nova.virt.hardware [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 994.911145] env[61970]: DEBUG nova.virt.hardware [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 994.911517] env[61970]: DEBUG nova.virt.hardware [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 994.911822] env[61970]: DEBUG nova.virt.hardware [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 994.912135] env[61970]: DEBUG nova.virt.hardware [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 994.912443] env[61970]: DEBUG nova.virt.hardware [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 994.913809] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d975db0-369a-434a-8017-b26b32d2e429 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.924874] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5978c35d-6f3b-42a8-b2b4-708d4c84ed4b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.973780] env[61970]: DEBUG oslo_vmware.api [None req-4f2e7e03-53c0-490a-b358-6cfcc0485979 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356010, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.023910] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 995.023910] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529ee39b-aaae-f730-9303-b5cdbb96fa41" [ 995.023910] env[61970]: _type = "HttpNfcLease" [ 995.023910] env[61970]: } is ready. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 995.024234] env[61970]: DEBUG oslo_vmware.rw_handles [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 995.024234] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]529ee39b-aaae-f730-9303-b5cdbb96fa41" [ 995.024234] env[61970]: _type = "HttpNfcLease" [ 995.024234] env[61970]: }. {{(pid=61970) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 995.024975] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e08803e-253c-4f3d-8c21-7c1ef2fc7353 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.032209] env[61970]: DEBUG oslo_vmware.rw_handles [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52607458-267d-232b-e967-cdd2cbe47b34/disk-0.vmdk from lease info. {{(pid=61970) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 995.032395] env[61970]: DEBUG oslo_vmware.rw_handles [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Creating HTTP connection to write to file with size = 31670784 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52607458-267d-232b-e967-cdd2cbe47b34/disk-0.vmdk. {{(pid=61970) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 995.121429] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0b0c8761-8850-40d8-9585-8a814fd7a91d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.152159] env[61970]: DEBUG oslo_vmware.api [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356003, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.152900] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4f6833d8-c011-4393-a487-e69c200fb1e8 tempest-ServerDiskConfigTestJSON-1052088114 tempest-ServerDiskConfigTestJSON-1052088114-project-member] Lock "3b1e8446-7862-4532-b2ed-611b63dbf391" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.025s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.347092] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356011, 'name': CreateVM_Task, 'duration_secs': 0.494089} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.347092] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 995.347389] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.347537] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.347866] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 995.348165] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-899994b3-e769-4d10-8370-70c5b571989f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.355647] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 995.355647] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a3c5ac-7db9-750e-4b08-68a8ed50a625" [ 995.355647] env[61970]: _type = "Task" [ 995.355647] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.363623] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a3c5ac-7db9-750e-4b08-68a8ed50a625, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.460430] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf674a9-ad16-4540-bdbb-4644cec1eb1a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.468476] env[61970]: DEBUG nova.compute.manager [req-1ecb4e7c-be24-4df1-968b-4f9a97829410 req-8a472e92-a1a3-4634-ae40-9548aab27ca5 service nova] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Received event network-vif-plugged-62d22328-b863-41dc-816f-1ce3f7a53eed {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.468888] env[61970]: DEBUG oslo_concurrency.lockutils [req-1ecb4e7c-be24-4df1-968b-4f9a97829410 req-8a472e92-a1a3-4634-ae40-9548aab27ca5 service nova] Acquiring lock "d38bde37-994b-4d58-b4b1-5f79b59fd2aa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.469041] env[61970]: DEBUG oslo_concurrency.lockutils [req-1ecb4e7c-be24-4df1-968b-4f9a97829410 req-8a472e92-a1a3-4634-ae40-9548aab27ca5 service nova] Lock "d38bde37-994b-4d58-b4b1-5f79b59fd2aa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.469317] env[61970]: DEBUG oslo_concurrency.lockutils [req-1ecb4e7c-be24-4df1-968b-4f9a97829410 req-8a472e92-a1a3-4634-ae40-9548aab27ca5 service nova] Lock "d38bde37-994b-4d58-b4b1-5f79b59fd2aa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.469550] env[61970]: DEBUG nova.compute.manager [req-1ecb4e7c-be24-4df1-968b-4f9a97829410 req-8a472e92-a1a3-4634-ae40-9548aab27ca5 service nova] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] No waiting events found dispatching network-vif-plugged-62d22328-b863-41dc-816f-1ce3f7a53eed {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 995.469790] env[61970]: WARNING nova.compute.manager [req-1ecb4e7c-be24-4df1-968b-4f9a97829410 req-8a472e92-a1a3-4634-ae40-9548aab27ca5 service nova] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Received unexpected event network-vif-plugged-62d22328-b863-41dc-816f-1ce3f7a53eed for instance with vm_state building and task_state spawning. [ 995.485604] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c121e7-1015-462a-ad67-a764ba129ce2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.494279] env[61970]: DEBUG oslo_vmware.api [None req-4f2e7e03-53c0-490a-b358-6cfcc0485979 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356010, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.538191] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e07a9d3-7ba6-4186-a8e9-c0a0dd86ec39 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.556586] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d8d0f44-3c89-48a3-a0dc-92f6f4d6dc5f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.585318] env[61970]: DEBUG nova.compute.provider_tree [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.635304] env[61970]: DEBUG nova.network.neutron [req-3998eab5-884c-4140-87d3-418f3fdbaf01 req-b9188a7d-536e-40a7-878f-ce7aab9c059d service nova] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Updated VIF entry in instance network info cache for port f7943b8e-3fab-4cfc-b59d-5a7a01efb9ba. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 995.635504] env[61970]: DEBUG nova.network.neutron [req-3998eab5-884c-4140-87d3-418f3fdbaf01 req-b9188a7d-536e-40a7-878f-ce7aab9c059d service nova] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Updating instance_info_cache with network_info: [{"id": "f7943b8e-3fab-4cfc-b59d-5a7a01efb9ba", "address": "fa:16:3e:70:b8:6e", "network": {"id": "e02bfa6c-8f0d-4413-8dcf-e644c07792d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1410864732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9235eeeb6fa4ccf959ffbea456f3694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7943b8e-3f", "ovs_interfaceid": "f7943b8e-3fab-4cfc-b59d-5a7a01efb9ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.652328] env[61970]: DEBUG oslo_vmware.api [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356003, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.869557] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a3c5ac-7db9-750e-4b08-68a8ed50a625, 'name': SearchDatastore_Task, 'duration_secs': 0.016133} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.870019] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.870345] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 995.870728] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.870942] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.870987] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 995.873218] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34522366-97df-4fa1-b1de-7724db29f9e9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.883462] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 995.883747] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 995.884796] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2290e6f3-4ad1-4ec6-8b81-a5ded9df65d3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.890788] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 995.890788] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d21178-1a88-d9f2-f20d-dbc046688295" [ 995.890788] env[61970]: _type = "Task" [ 995.890788] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.901161] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d21178-1a88-d9f2-f20d-dbc046688295, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.978525] env[61970]: DEBUG oslo_vmware.api [None req-4f2e7e03-53c0-490a-b358-6cfcc0485979 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356010, 'name': PowerOffVM_Task, 'duration_secs': 1.24311} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.978902] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f2e7e03-53c0-490a-b358-6cfcc0485979 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 995.979128] env[61970]: DEBUG nova.compute.manager [None req-4f2e7e03-53c0-490a-b358-6cfcc0485979 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 995.980303] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8610c4f4-a98f-4830-bd5c-4a7a70056236 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.058530] env[61970]: DEBUG nova.network.neutron [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Successfully updated port: 62d22328-b863-41dc-816f-1ce3f7a53eed {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 996.082108] env[61970]: DEBUG nova.compute.manager [req-031cfbee-8b71-4ea4-9ff1-0c670053743a req-2eab1979-3ed6-4cbe-9ab8-87d77b2511b9 service nova] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Received event network-changed-62d22328-b863-41dc-816f-1ce3f7a53eed {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 996.082335] env[61970]: DEBUG nova.compute.manager [req-031cfbee-8b71-4ea4-9ff1-0c670053743a req-2eab1979-3ed6-4cbe-9ab8-87d77b2511b9 service nova] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Refreshing instance network info cache due to event network-changed-62d22328-b863-41dc-816f-1ce3f7a53eed. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 996.082555] env[61970]: DEBUG oslo_concurrency.lockutils [req-031cfbee-8b71-4ea4-9ff1-0c670053743a req-2eab1979-3ed6-4cbe-9ab8-87d77b2511b9 service nova] Acquiring lock "refresh_cache-d38bde37-994b-4d58-b4b1-5f79b59fd2aa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.082783] env[61970]: DEBUG oslo_concurrency.lockutils [req-031cfbee-8b71-4ea4-9ff1-0c670053743a req-2eab1979-3ed6-4cbe-9ab8-87d77b2511b9 service nova] Acquired lock "refresh_cache-d38bde37-994b-4d58-b4b1-5f79b59fd2aa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.083017] env[61970]: DEBUG nova.network.neutron [req-031cfbee-8b71-4ea4-9ff1-0c670053743a req-2eab1979-3ed6-4cbe-9ab8-87d77b2511b9 service nova] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Refreshing network info cache for port 62d22328-b863-41dc-816f-1ce3f7a53eed {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 996.098101] env[61970]: DEBUG nova.scheduler.client.report [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.138595] env[61970]: DEBUG oslo_concurrency.lockutils [req-3998eab5-884c-4140-87d3-418f3fdbaf01 req-b9188a7d-536e-40a7-878f-ce7aab9c059d service nova] Releasing lock "refresh_cache-00e3287d-24fb-4ea9-b212-54215828d5df" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.139111] env[61970]: DEBUG nova.compute.manager [req-3998eab5-884c-4140-87d3-418f3fdbaf01 req-b9188a7d-536e-40a7-878f-ce7aab9c059d service nova] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Received event network-vif-deleted-7bfd07fd-97bc-42b5-9571-0807d864f9bc {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 996.152487] env[61970]: DEBUG oslo_vmware.api [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356003, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.246983] env[61970]: DEBUG nova.compute.manager [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 996.247974] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd780786-a77f-4ba9-a21a-cecc4224d03a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.342193] env[61970]: DEBUG oslo_vmware.rw_handles [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Completed reading data from the image iterator. {{(pid=61970) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 996.342400] env[61970]: DEBUG oslo_vmware.rw_handles [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52607458-267d-232b-e967-cdd2cbe47b34/disk-0.vmdk. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 996.343404] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32196729-6df0-42fd-a8d5-7e61c6c3c3be {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.352169] env[61970]: DEBUG oslo_vmware.rw_handles [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52607458-267d-232b-e967-cdd2cbe47b34/disk-0.vmdk is in state: ready. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 996.352376] env[61970]: DEBUG oslo_vmware.rw_handles [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52607458-267d-232b-e967-cdd2cbe47b34/disk-0.vmdk. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 996.354666] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-fae3568a-f37d-409b-93ee-c8c46c117962 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.408926] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d21178-1a88-d9f2-f20d-dbc046688295, 'name': SearchDatastore_Task, 'duration_secs': 0.00972} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.410486] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1921a4de-bacb-4634-b150-44ee93f64c34 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.415373] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 996.415373] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5274afb0-11d0-1a9a-9901-25bcdbd92ff7" [ 996.415373] env[61970]: _type = "Task" [ 996.415373] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.423401] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5274afb0-11d0-1a9a-9901-25bcdbd92ff7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.491406] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4f2e7e03-53c0-490a-b358-6cfcc0485979 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.557s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.519828] env[61970]: DEBUG oslo_vmware.rw_handles [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52607458-267d-232b-e967-cdd2cbe47b34/disk-0.vmdk. {{(pid=61970) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 996.520091] env[61970]: INFO nova.virt.vmwareapi.images [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Downloaded image file data 8ff34e61-3203-40b2-805a-fd5355a36d65 [ 996.521031] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4aefd2b-c9c1-40e4-aa5b-41cab6fc5f63 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.536969] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-25f13895-ac42-4b2b-a121-67e5f4888b72 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.561821] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "refresh_cache-d38bde37-994b-4d58-b4b1-5f79b59fd2aa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.576980] env[61970]: INFO nova.virt.vmwareapi.images [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] The imported VM was unregistered [ 996.579543] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Caching image {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 996.579798] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Creating directory with path [datastore2] devstack-image-cache_base/8ff34e61-3203-40b2-805a-fd5355a36d65 {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 996.580241] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8b42e6df-5411-4515-9601-709c0802df83 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.591116] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Created directory with path [datastore2] devstack-image-cache_base/8ff34e61-3203-40b2-805a-fd5355a36d65 {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 996.591319] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_c0a6eb66-edfc-4d4a-9b6a-ccd264c124b8/OSTACK_IMG_c0a6eb66-edfc-4d4a-9b6a-ccd264c124b8.vmdk to [datastore2] devstack-image-cache_base/8ff34e61-3203-40b2-805a-fd5355a36d65/8ff34e61-3203-40b2-805a-fd5355a36d65.vmdk. {{(pid=61970) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 996.591892] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-fe106945-70da-4958-b45d-1b589a34021d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.598609] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 996.598609] env[61970]: value = "task-1356013" [ 996.598609] env[61970]: _type = "Task" [ 996.598609] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.606302] env[61970]: DEBUG oslo_concurrency.lockutils [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.913s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.608077] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356013, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.627985] env[61970]: DEBUG nova.network.neutron [req-031cfbee-8b71-4ea4-9ff1-0c670053743a req-2eab1979-3ed6-4cbe-9ab8-87d77b2511b9 service nova] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 996.630738] env[61970]: INFO nova.scheduler.client.report [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Deleted allocations for instance 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa [ 996.653720] env[61970]: DEBUG oslo_vmware.api [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356003, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.759647] env[61970]: INFO nova.compute.manager [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] instance snapshotting [ 996.760346] env[61970]: DEBUG nova.objects.instance [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lazy-loading 'flavor' on Instance uuid ebb92a5b-f635-4504-a57f-395de514015a {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.765187] env[61970]: DEBUG nova.network.neutron [req-031cfbee-8b71-4ea4-9ff1-0c670053743a req-2eab1979-3ed6-4cbe-9ab8-87d77b2511b9 service nova] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.926396] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5274afb0-11d0-1a9a-9901-25bcdbd92ff7, 'name': SearchDatastore_Task, 'duration_secs': 0.008739} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.926833] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.926982] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 00e3287d-24fb-4ea9-b212-54215828d5df/00e3287d-24fb-4ea9-b212-54215828d5df.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 996.927287] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b6202b9-3154-4309-b1fc-2b712a1a58c6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.937753] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 996.937753] env[61970]: value = "task-1356014" [ 996.937753] env[61970]: _type = "Task" [ 996.937753] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.946456] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356014, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.108646] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356013, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.141042] env[61970]: DEBUG oslo_concurrency.lockutils [None req-557f65a6-97bc-4482-8aeb-25f6b0e2a0a0 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.770s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.156782] env[61970]: DEBUG oslo_vmware.api [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356003, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.266409] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e9a762-a8ab-48a9-942d-97fb2840d5e4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.270623] env[61970]: DEBUG oslo_concurrency.lockutils [req-031cfbee-8b71-4ea4-9ff1-0c670053743a req-2eab1979-3ed6-4cbe-9ab8-87d77b2511b9 service nova] Releasing lock "refresh_cache-d38bde37-994b-4d58-b4b1-5f79b59fd2aa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.271013] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "refresh_cache-d38bde37-994b-4d58-b4b1-5f79b59fd2aa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.271191] env[61970]: DEBUG nova.network.neutron [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 997.294589] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-752c00ac-d099-4b94-b60e-4993883aa0c8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.374739] env[61970]: DEBUG nova.objects.instance [None req-674eab95-37c2-4159-ac47-adcb25e91067 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lazy-loading 'flavor' on Instance uuid d0ad9689-bdb7-4d68-b19a-3f805ab4612e {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.448779] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356014, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.612213] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356013, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.657742] env[61970]: DEBUG oslo_vmware.api [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356003, 'name': ReconfigVM_Task, 'duration_secs': 5.801449} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.658221] env[61970]: DEBUG oslo_concurrency.lockutils [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.658702] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Reconfigured VM to detach interface {{(pid=61970) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 997.809020] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Creating Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 997.809020] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b0a284ef-83cf-4e65-9237-da793d625c5c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.815858] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 997.815858] env[61970]: value = "task-1356015" [ 997.815858] env[61970]: _type = "Task" [ 997.815858] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.831138] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356015, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.832522] env[61970]: DEBUG nova.network.neutron [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 997.884464] env[61970]: DEBUG oslo_concurrency.lockutils [None req-674eab95-37c2-4159-ac47-adcb25e91067 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "refresh_cache-d0ad9689-bdb7-4d68-b19a-3f805ab4612e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.884680] env[61970]: DEBUG oslo_concurrency.lockutils [None req-674eab95-37c2-4159-ac47-adcb25e91067 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquired lock "refresh_cache-d0ad9689-bdb7-4d68-b19a-3f805ab4612e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.884865] env[61970]: DEBUG nova.network.neutron [None req-674eab95-37c2-4159-ac47-adcb25e91067 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 997.885061] env[61970]: DEBUG nova.objects.instance [None req-674eab95-37c2-4159-ac47-adcb25e91067 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lazy-loading 'info_cache' on Instance uuid d0ad9689-bdb7-4d68-b19a-3f805ab4612e {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.949979] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356014, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.110721] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356013, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.182613] env[61970]: DEBUG nova.network.neutron [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Updating instance_info_cache with network_info: [{"id": "62d22328-b863-41dc-816f-1ce3f7a53eed", "address": "fa:16:3e:53:70:05", "network": {"id": "62edd125-b77c-4418-a465-1dcb3a0733c1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-371307366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6a4997482a4834a30c79c4cbab2cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62d22328-b8", "ovs_interfaceid": "62d22328-b863-41dc-816f-1ce3f7a53eed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.330097] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356015, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.390660] env[61970]: DEBUG nova.objects.base [None req-674eab95-37c2-4159-ac47-adcb25e91067 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 998.460083] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356014, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.611068] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356013, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.688039] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "refresh_cache-d38bde37-994b-4d58-b4b1-5f79b59fd2aa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.688039] env[61970]: DEBUG nova.compute.manager [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Instance network_info: |[{"id": "62d22328-b863-41dc-816f-1ce3f7a53eed", "address": "fa:16:3e:53:70:05", "network": {"id": "62edd125-b77c-4418-a465-1dcb3a0733c1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-371307366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6a4997482a4834a30c79c4cbab2cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62d22328-b8", "ovs_interfaceid": "62d22328-b863-41dc-816f-1ce3f7a53eed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 998.688039] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:70:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '62d22328-b863-41dc-816f-1ce3f7a53eed', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 998.696736] env[61970]: DEBUG oslo.service.loopingcall [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 998.697151] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 998.697523] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-25c61ebc-1cb7-413a-895c-c256810fc033 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.720708] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 998.720708] env[61970]: value = "task-1356016" [ 998.720708] env[61970]: _type = "Task" [ 998.720708] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.728914] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356016, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.828224] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356015, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.955018] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356014, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.112942] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356013, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.235167] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356016, 'name': CreateVM_Task} progress is 25%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.328684] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356015, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.416488] env[61970]: DEBUG nova.network.neutron [None req-674eab95-37c2-4159-ac47-adcb25e91067 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Updating instance_info_cache with network_info: [{"id": "b83024be-4171-40cd-9d1c-65a2d2ce8b9e", "address": "fa:16:3e:f5:18:c9", "network": {"id": "1f53a96b-8a46-41d5-8bf3-759de1fce443", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-108385347-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a44c05a894394d69b8b4fd36c7b2532a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb83024be-41", "ovs_interfaceid": "b83024be-4171-40cd-9d1c-65a2d2ce8b9e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.432052] env[61970]: DEBUG oslo_concurrency.lockutils [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.432052] env[61970]: DEBUG oslo_concurrency.lockutils [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.432052] env[61970]: DEBUG nova.network.neutron [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 999.451211] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356014, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.416287} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.451481] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 00e3287d-24fb-4ea9-b212-54215828d5df/00e3287d-24fb-4ea9-b212-54215828d5df.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 999.451696] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 999.451943] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b30d2e21-ac55-4d51-a5a3-c3b7a637ac86 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.460274] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 999.460274] env[61970]: value = "task-1356017" [ 999.460274] env[61970]: _type = "Task" [ 999.460274] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.469615] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356017, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.612035] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356013, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.796848} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.612326] env[61970]: INFO nova.virt.vmwareapi.ds_util [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_c0a6eb66-edfc-4d4a-9b6a-ccd264c124b8/OSTACK_IMG_c0a6eb66-edfc-4d4a-9b6a-ccd264c124b8.vmdk to [datastore2] devstack-image-cache_base/8ff34e61-3203-40b2-805a-fd5355a36d65/8ff34e61-3203-40b2-805a-fd5355a36d65.vmdk. [ 999.612512] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Cleaning up location [datastore2] OSTACK_IMG_c0a6eb66-edfc-4d4a-9b6a-ccd264c124b8 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 999.612672] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_c0a6eb66-edfc-4d4a-9b6a-ccd264c124b8 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 999.612927] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fbff3043-96be-44f2-97f0-1d2e0aae53c7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.620310] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 999.620310] env[61970]: value = "task-1356018" [ 999.620310] env[61970]: _type = "Task" [ 999.620310] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.631514] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356018, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.731985] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356016, 'name': CreateVM_Task, 'duration_secs': 0.878656} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.733107] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 999.733337] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.733472] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.733809] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 999.734085] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70acd9e5-5f07-4fbd-947b-02e513d93324 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.742210] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 999.742210] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52dbe9da-6476-3c7f-503a-8bffddb250fd" [ 999.742210] env[61970]: _type = "Task" [ 999.742210] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.752749] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52dbe9da-6476-3c7f-503a-8bffddb250fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.831106] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356015, 'name': CreateSnapshot_Task, 'duration_secs': 1.513714} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.831106] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Created Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 999.831106] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d50ba95-7830-481c-9408-e83a20ccfd36 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.919894] env[61970]: DEBUG oslo_concurrency.lockutils [None req-674eab95-37c2-4159-ac47-adcb25e91067 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Releasing lock "refresh_cache-d0ad9689-bdb7-4d68-b19a-3f805ab4612e" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.975502] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356017, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06737} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.978811] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 999.978811] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-449d29fb-e8a5-4e5b-a02b-60869072a4fc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.006018] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 00e3287d-24fb-4ea9-b212-54215828d5df/00e3287d-24fb-4ea9-b212-54215828d5df.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1000.006018] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-acbb6e36-f7ad-41fd-9f66-c265cf0e02c6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.028110] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 1000.028110] env[61970]: value = "task-1356019" [ 1000.028110] env[61970]: _type = "Task" [ 1000.028110] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.039161] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356019, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.130396] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356018, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.032866} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.130698] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1000.130985] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8ff34e61-3203-40b2-805a-fd5355a36d65/8ff34e61-3203-40b2-805a-fd5355a36d65.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.131155] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8ff34e61-3203-40b2-805a-fd5355a36d65/8ff34e61-3203-40b2-805a-fd5355a36d65.vmdk to [datastore2] 7f59000a-94d9-45b6-aa7a-300d95793615/7f59000a-94d9-45b6-aa7a-300d95793615.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1000.131533] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d6e45bbb-68aa-4092-a7e1-954d6dd59bc0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.138467] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1000.138467] env[61970]: value = "task-1356020" [ 1000.138467] env[61970]: _type = "Task" [ 1000.138467] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.149451] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356020, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.212039] env[61970]: INFO nova.network.neutron [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Port 90fbe562-a7de-4ce8-a337-3c3f7be95120 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1000.213280] env[61970]: DEBUG nova.network.neutron [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Updating instance_info_cache with network_info: [{"id": "5b506f12-7feb-41e2-8e18-abc9262ac87b", "address": "fa:16:3e:f9:43:05", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b506f12-7f", "ovs_interfaceid": "5b506f12-7feb-41e2-8e18-abc9262ac87b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.257198] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52dbe9da-6476-3c7f-503a-8bffddb250fd, 'name': SearchDatastore_Task, 'duration_secs': 0.009494} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.257198] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.257790] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1000.263181] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.263181] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.263181] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1000.263181] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9116f2de-21af-460b-a4e0-3f33be725c10 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.263999] env[61970]: DEBUG nova.compute.manager [req-2965d79d-4087-4d42-986c-2f3ed9ad1690 req-ad95587e-8ed9-4d8f-ba02-629746ec862e service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Received event network-changed-5b506f12-7feb-41e2-8e18-abc9262ac87b {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.263999] env[61970]: DEBUG nova.compute.manager [req-2965d79d-4087-4d42-986c-2f3ed9ad1690 req-ad95587e-8ed9-4d8f-ba02-629746ec862e service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Refreshing instance network info cache due to event network-changed-5b506f12-7feb-41e2-8e18-abc9262ac87b. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1000.264288] env[61970]: DEBUG oslo_concurrency.lockutils [req-2965d79d-4087-4d42-986c-2f3ed9ad1690 req-ad95587e-8ed9-4d8f-ba02-629746ec862e service nova] Acquiring lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.271085] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1000.271085] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1000.271571] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc1df4f7-7e27-42fe-9b48-a61a97375e51 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.286823] env[61970]: DEBUG oslo_concurrency.lockutils [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "31f9d37b-f4ff-4a8b-9477-8253f8305020" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.286823] env[61970]: DEBUG oslo_concurrency.lockutils [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "31f9d37b-f4ff-4a8b-9477-8253f8305020" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.287215] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 1000.287215] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]521cc1bf-52f1-f3a7-b35d-33d112284af3" [ 1000.287215] env[61970]: _type = "Task" [ 1000.287215] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.300673] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]521cc1bf-52f1-f3a7-b35d-33d112284af3, 'name': SearchDatastore_Task, 'duration_secs': 0.012555} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.301993] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d38a4875-030c-4063-8a21-d9d538b537b6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.311339] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 1000.311339] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525b2635-27dc-cba4-4c33-c0b1f1aaacad" [ 1000.311339] env[61970]: _type = "Task" [ 1000.311339] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.321050] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525b2635-27dc-cba4-4c33-c0b1f1aaacad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.349476] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Creating linked-clone VM from snapshot {{(pid=61970) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1000.349768] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-db698b5a-862d-4dfa-ac5f-3ac2ad3380cb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.359371] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1000.359371] env[61970]: value = "task-1356021" [ 1000.359371] env[61970]: _type = "Task" [ 1000.359371] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.368835] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356021, 'name': CloneVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.423311] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-674eab95-37c2-4159-ac47-adcb25e91067 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1000.423709] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07fe3500-4216-4b63-b8ca-15d097a93aaa {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.432112] env[61970]: DEBUG oslo_vmware.api [None req-674eab95-37c2-4159-ac47-adcb25e91067 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1000.432112] env[61970]: value = "task-1356022" [ 1000.432112] env[61970]: _type = "Task" [ 1000.432112] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.443390] env[61970]: DEBUG oslo_vmware.api [None req-674eab95-37c2-4159-ac47-adcb25e91067 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356022, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.540491] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356019, 'name': ReconfigVM_Task, 'duration_secs': 0.299102} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.540872] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 00e3287d-24fb-4ea9-b212-54215828d5df/00e3287d-24fb-4ea9-b212-54215828d5df.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1000.541723] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-adba662e-de93-46bc-97c1-a9c8f09e75ac {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.549812] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 1000.549812] env[61970]: value = "task-1356023" [ 1000.549812] env[61970]: _type = "Task" [ 1000.549812] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.558652] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356023, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.652085] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356020, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.715391] env[61970]: DEBUG oslo_concurrency.lockutils [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.723021] env[61970]: DEBUG oslo_concurrency.lockutils [req-2965d79d-4087-4d42-986c-2f3ed9ad1690 req-ad95587e-8ed9-4d8f-ba02-629746ec862e service nova] Acquired lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.723021] env[61970]: DEBUG nova.network.neutron [req-2965d79d-4087-4d42-986c-2f3ed9ad1690 req-ad95587e-8ed9-4d8f-ba02-629746ec862e service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Refreshing network info cache for port 5b506f12-7feb-41e2-8e18-abc9262ac87b {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1000.787884] env[61970]: DEBUG nova.compute.manager [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1000.824804] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525b2635-27dc-cba4-4c33-c0b1f1aaacad, 'name': SearchDatastore_Task, 'duration_secs': 0.083712} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.825098] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.825375] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] d38bde37-994b-4d58-b4b1-5f79b59fd2aa/d38bde37-994b-4d58-b4b1-5f79b59fd2aa.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1000.825747] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6666f3c2-4290-4454-a3e4-9c8fa0d82c9a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.833779] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 1000.833779] env[61970]: value = "task-1356024" [ 1000.833779] env[61970]: _type = "Task" [ 1000.833779] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.843691] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356024, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.869775] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356021, 'name': CloneVM_Task} progress is 94%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.946662] env[61970]: DEBUG oslo_vmware.api [None req-674eab95-37c2-4159-ac47-adcb25e91067 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356022, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.013604] env[61970]: DEBUG oslo_concurrency.lockutils [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "interface-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8-90fbe562-a7de-4ce8-a337-3c3f7be95120" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.013604] env[61970]: DEBUG oslo_concurrency.lockutils [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "interface-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8-90fbe562-a7de-4ce8-a337-3c3f7be95120" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.013604] env[61970]: DEBUG nova.objects.instance [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lazy-loading 'flavor' on Instance uuid c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.061974] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356023, 'name': Rename_Task, 'duration_secs': 0.434391} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.062333] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1001.062669] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-47a0fb72-9669-4bb1-8021-892c085f2440 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.071120] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 1001.071120] env[61970]: value = "task-1356025" [ 1001.071120] env[61970]: _type = "Task" [ 1001.071120] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.082514] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356025, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.150488] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356020, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.220195] env[61970]: DEBUG oslo_concurrency.lockutils [None req-613b7599-e902-4ffb-ae2c-9a1223e17aef tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "interface-c4ec1797-d939-4c24-9314-ce9c74eb5aa1-90fbe562-a7de-4ce8-a337-3c3f7be95120" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.178s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.322356] env[61970]: DEBUG oslo_concurrency.lockutils [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.322676] env[61970]: DEBUG oslo_concurrency.lockutils [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.324561] env[61970]: INFO nova.compute.claims [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1001.345446] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356024, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.372610] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356021, 'name': CloneVM_Task} progress is 94%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.446204] env[61970]: DEBUG oslo_vmware.api [None req-674eab95-37c2-4159-ac47-adcb25e91067 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356022, 'name': PowerOnVM_Task, 'duration_secs': 0.888752} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.446732] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-674eab95-37c2-4159-ac47-adcb25e91067 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1001.447060] env[61970]: DEBUG nova.compute.manager [None req-674eab95-37c2-4159-ac47-adcb25e91067 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1001.448265] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ce67eca-b242-4d0f-9927-6a8d34e646c5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.522539] env[61970]: DEBUG nova.network.neutron [req-2965d79d-4087-4d42-986c-2f3ed9ad1690 req-ad95587e-8ed9-4d8f-ba02-629746ec862e service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Updated VIF entry in instance network info cache for port 5b506f12-7feb-41e2-8e18-abc9262ac87b. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1001.523089] env[61970]: DEBUG nova.network.neutron [req-2965d79d-4087-4d42-986c-2f3ed9ad1690 req-ad95587e-8ed9-4d8f-ba02-629746ec862e service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Updating instance_info_cache with network_info: [{"id": "5b506f12-7feb-41e2-8e18-abc9262ac87b", "address": "fa:16:3e:f9:43:05", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b506f12-7f", "ovs_interfaceid": "5b506f12-7feb-41e2-8e18-abc9262ac87b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.583867] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356025, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.651614] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356020, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.683349] env[61970]: DEBUG nova.objects.instance [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lazy-loading 'pci_requests' on Instance uuid c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.852371] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356024, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.871587] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356021, 'name': CloneVM_Task} progress is 94%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.026260] env[61970]: DEBUG oslo_concurrency.lockutils [req-2965d79d-4087-4d42-986c-2f3ed9ad1690 req-ad95587e-8ed9-4d8f-ba02-629746ec862e service nova] Releasing lock "refresh_cache-c4ec1797-d939-4c24-9314-ce9c74eb5aa1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.085388] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356025, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.152922] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356020, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.188970] env[61970]: DEBUG nova.objects.base [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1002.189134] env[61970]: DEBUG nova.network.neutron [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1002.317634] env[61970]: DEBUG nova.policy [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2af915dc43c84940a937dfb8a04ea0b6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17ca05a1e6664430a02de563d98c1148', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 1002.349528] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356024, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.372645] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356021, 'name': CloneVM_Task} progress is 94%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.544534] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29f64249-8756-4da8-8303-8135d2edff38 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.552465] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1455e42f-8522-45c5-8a22-e64789dfeecd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.588230] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73feeb37-0fa1-49f7-bb4b-3f7a14687972 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.598304] env[61970]: DEBUG oslo_vmware.api [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356025, 'name': PowerOnVM_Task, 'duration_secs': 1.23773} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.600572] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1002.601500] env[61970]: INFO nova.compute.manager [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Took 10.02 seconds to spawn the instance on the hypervisor. [ 1002.601718] env[61970]: DEBUG nova.compute.manager [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1002.603053] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d40d9bbd-2a69-48bb-802c-6c7da371d996 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.606280] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa1d3c9-7d81-4f4a-b057-2bc734d0400a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.621712] env[61970]: DEBUG nova.compute.provider_tree [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.649447] env[61970]: DEBUG nova.compute.manager [req-6b99caac-7555-4e8f-83e7-139826418869 req-71bce187-7a67-4026-90b9-56c4573faf57 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Received event network-changed-4434695d-ddda-4840-b1da-8159c98dcae0 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.649687] env[61970]: DEBUG nova.compute.manager [req-6b99caac-7555-4e8f-83e7-139826418869 req-71bce187-7a67-4026-90b9-56c4573faf57 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Refreshing instance network info cache due to event network-changed-4434695d-ddda-4840-b1da-8159c98dcae0. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1002.649910] env[61970]: DEBUG oslo_concurrency.lockutils [req-6b99caac-7555-4e8f-83e7-139826418869 req-71bce187-7a67-4026-90b9-56c4573faf57 service nova] Acquiring lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.650075] env[61970]: DEBUG oslo_concurrency.lockutils [req-6b99caac-7555-4e8f-83e7-139826418869 req-71bce187-7a67-4026-90b9-56c4573faf57 service nova] Acquired lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.650243] env[61970]: DEBUG nova.network.neutron [req-6b99caac-7555-4e8f-83e7-139826418869 req-71bce187-7a67-4026-90b9-56c4573faf57 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Refreshing network info cache for port 4434695d-ddda-4840-b1da-8159c98dcae0 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1002.658733] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356020, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.463345} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.659352] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8ff34e61-3203-40b2-805a-fd5355a36d65/8ff34e61-3203-40b2-805a-fd5355a36d65.vmdk to [datastore2] 7f59000a-94d9-45b6-aa7a-300d95793615/7f59000a-94d9-45b6-aa7a-300d95793615.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1002.660304] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9ad581e-0055-48fc-af85-e9e970909d42 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.690277] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 7f59000a-94d9-45b6-aa7a-300d95793615/7f59000a-94d9-45b6-aa7a-300d95793615.vmdk or device None with type streamOptimized {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1002.692043] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aac4a3ea-25dd-4d0b-8d80-86aa1cf4078e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.716112] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1002.716112] env[61970]: value = "task-1356026" [ 1002.716112] env[61970]: _type = "Task" [ 1002.716112] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.725825] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356026, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.772876] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Acquiring lock "19c0bfcb-1964-4719-98b1-d9334751ace2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.773151] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Lock "19c0bfcb-1964-4719-98b1-d9334751ace2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.847698] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356024, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.876469] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356021, 'name': CloneVM_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.130993] env[61970]: DEBUG nova.scheduler.client.report [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.144524] env[61970]: INFO nova.compute.manager [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Took 19.25 seconds to build instance. [ 1003.229902] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356026, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.276012] env[61970]: DEBUG nova.compute.manager [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1003.347858] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356024, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.08556} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.348561] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] d38bde37-994b-4d58-b4b1-5f79b59fd2aa/d38bde37-994b-4d58-b4b1-5f79b59fd2aa.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1003.348822] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1003.349147] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df0aba4d-4fcf-46d8-adfd-82191edb7a47 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.355920] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 1003.355920] env[61970]: value = "task-1356027" [ 1003.355920] env[61970]: _type = "Task" [ 1003.355920] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.364367] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356027, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.375616] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356021, 'name': CloneVM_Task, 'duration_secs': 2.527654} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.375765] env[61970]: INFO nova.virt.vmwareapi.vmops [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Created linked-clone VM from snapshot [ 1003.377224] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df448996-e6db-417f-85cf-da3a8df5c7ab {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.384644] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Uploading image d497d9c2-91aa-45aa-90d9-3927851fe30e {{(pid=61970) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1003.407793] env[61970]: DEBUG nova.network.neutron [req-6b99caac-7555-4e8f-83e7-139826418869 req-71bce187-7a67-4026-90b9-56c4573faf57 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Updated VIF entry in instance network info cache for port 4434695d-ddda-4840-b1da-8159c98dcae0. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1003.408294] env[61970]: DEBUG nova.network.neutron [req-6b99caac-7555-4e8f-83e7-139826418869 req-71bce187-7a67-4026-90b9-56c4573faf57 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Updating instance_info_cache with network_info: [{"id": "4434695d-ddda-4840-b1da-8159c98dcae0", "address": "fa:16:3e:85:27:97", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4434695d-dd", "ovs_interfaceid": "4434695d-ddda-4840-b1da-8159c98dcae0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.417024] env[61970]: DEBUG oslo_vmware.rw_handles [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1003.417024] env[61970]: value = "vm-288862" [ 1003.417024] env[61970]: _type = "VirtualMachine" [ 1003.417024] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1003.417024] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ab6585fd-bdf5-4124-b904-4df7dce65986 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.428301] env[61970]: DEBUG oslo_vmware.rw_handles [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lease: (returnval){ [ 1003.428301] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52881d84-6f3c-50a6-34e4-978a6018d4bd" [ 1003.428301] env[61970]: _type = "HttpNfcLease" [ 1003.428301] env[61970]: } obtained for exporting VM: (result){ [ 1003.428301] env[61970]: value = "vm-288862" [ 1003.428301] env[61970]: _type = "VirtualMachine" [ 1003.428301] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1003.428301] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the lease: (returnval){ [ 1003.428301] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52881d84-6f3c-50a6-34e4-978a6018d4bd" [ 1003.428301] env[61970]: _type = "HttpNfcLease" [ 1003.428301] env[61970]: } to be ready. {{(pid=61970) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1003.438088] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1003.438088] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52881d84-6f3c-50a6-34e4-978a6018d4bd" [ 1003.438088] env[61970]: _type = "HttpNfcLease" [ 1003.438088] env[61970]: } is ready. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1003.438667] env[61970]: DEBUG oslo_vmware.rw_handles [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1003.438667] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52881d84-6f3c-50a6-34e4-978a6018d4bd" [ 1003.438667] env[61970]: _type = "HttpNfcLease" [ 1003.438667] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1003.439653] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7599add8-0a71-440c-aa52-6bc5466e0c82 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.448074] env[61970]: DEBUG oslo_vmware.rw_handles [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52552c28-61b0-83ac-075d-879f4acc9ebc/disk-0.vmdk from lease info. {{(pid=61970) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1003.448478] env[61970]: DEBUG oslo_vmware.rw_handles [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52552c28-61b0-83ac-075d-879f4acc9ebc/disk-0.vmdk for reading. {{(pid=61970) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1003.553494] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-cd86f4cc-267a-467e-b84a-650d18050ac2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.641234] env[61970]: DEBUG oslo_concurrency.lockutils [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.318s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.641859] env[61970]: DEBUG nova.compute.manager [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1003.646385] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ad368b9b-e506-47d7-a9ae-a364ac178108 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "00e3287d-24fb-4ea9-b212-54215828d5df" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.761s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.728387] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356026, 'name': ReconfigVM_Task, 'duration_secs': 0.697634} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.729554] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 7f59000a-94d9-45b6-aa7a-300d95793615/7f59000a-94d9-45b6-aa7a-300d95793615.vmdk or device None with type streamOptimized {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1003.732404] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_name': '/dev/sda', 'encryption_secret_uuid': None, 'size': 0, 'encryption_options': None, 'device_type': 'disk', 'disk_bus': None, 'encrypted': False, 'guest_format': None, 'boot_index': 0, 'encryption_format': None, 'image_id': '7ff5af44-3f73-4c84-82e5-a1543db26d03'}], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': False, 'attachment_id': 'eedbcc21-7f74-484a-94dd-1a6c0645a263', 'boot_index': None, 'device_type': None, 'mount_device': '/dev/sdb', 'disk_bus': None, 'guest_format': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288853', 'volume_id': 'a78067a3-226b-4aec-bb5b-6fb37d5bfca3', 'name': 'volume-a78067a3-226b-4aec-bb5b-6fb37d5bfca3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '7f59000a-94d9-45b6-aa7a-300d95793615', 'attached_at': '', 'detached_at': '', 'volume_id': 'a78067a3-226b-4aec-bb5b-6fb37d5bfca3', 'serial': 'a78067a3-226b-4aec-bb5b-6fb37d5bfca3'}, 'volume_type': None}], 'swap': None} {{(pid=61970) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1003.732794] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Volume attach. Driver type: vmdk {{(pid=61970) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1003.733172] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288853', 'volume_id': 'a78067a3-226b-4aec-bb5b-6fb37d5bfca3', 'name': 'volume-a78067a3-226b-4aec-bb5b-6fb37d5bfca3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '7f59000a-94d9-45b6-aa7a-300d95793615', 'attached_at': '', 'detached_at': '', 'volume_id': 'a78067a3-226b-4aec-bb5b-6fb37d5bfca3', 'serial': 'a78067a3-226b-4aec-bb5b-6fb37d5bfca3'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1003.735362] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a3e1945-6db8-4bec-8ac4-f5471888d81e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.751723] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f94b3a-51c0-4e29-a5c2-78aaee8812b4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.779285] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] volume-a78067a3-226b-4aec-bb5b-6fb37d5bfca3/volume-a78067a3-226b-4aec-bb5b-6fb37d5bfca3.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1003.784574] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbdf93a2-ba7f-4269-bb3e-b3f046474fe2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.806181] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1003.806181] env[61970]: value = "task-1356029" [ 1003.806181] env[61970]: _type = "Task" [ 1003.806181] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.580974] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.581301] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.582863] env[61970]: INFO nova.compute.claims [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1004.588423] env[61970]: DEBUG nova.network.neutron [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Successfully updated port: 90fbe562-a7de-4ce8-a337-3c3f7be95120 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1004.589996] env[61970]: DEBUG oslo_concurrency.lockutils [req-6b99caac-7555-4e8f-83e7-139826418869 req-71bce187-7a67-4026-90b9-56c4573faf57 service nova] Releasing lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.591230] env[61970]: DEBUG nova.compute.utils [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1004.593244] env[61970]: DEBUG nova.compute.manager [req-f2619d2e-3a9a-467a-b11a-2918a7e0cafc req-a28f1bb4-58c1-4c18-b1a6-ef220e6079c4 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Received event network-vif-plugged-90fbe562-a7de-4ce8-a337-3c3f7be95120 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.593563] env[61970]: DEBUG oslo_concurrency.lockutils [req-f2619d2e-3a9a-467a-b11a-2918a7e0cafc req-a28f1bb4-58c1-4c18-b1a6-ef220e6079c4 service nova] Acquiring lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.593834] env[61970]: DEBUG oslo_concurrency.lockutils [req-f2619d2e-3a9a-467a-b11a-2918a7e0cafc req-a28f1bb4-58c1-4c18-b1a6-ef220e6079c4 service nova] Lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.594115] env[61970]: DEBUG oslo_concurrency.lockutils [req-f2619d2e-3a9a-467a-b11a-2918a7e0cafc req-a28f1bb4-58c1-4c18-b1a6-ef220e6079c4 service nova] Lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.594379] env[61970]: DEBUG nova.compute.manager [req-f2619d2e-3a9a-467a-b11a-2918a7e0cafc req-a28f1bb4-58c1-4c18-b1a6-ef220e6079c4 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] No waiting events found dispatching network-vif-plugged-90fbe562-a7de-4ce8-a337-3c3f7be95120 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1004.594616] env[61970]: WARNING nova.compute.manager [req-f2619d2e-3a9a-467a-b11a-2918a7e0cafc req-a28f1bb4-58c1-4c18-b1a6-ef220e6079c4 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Received unexpected event network-vif-plugged-90fbe562-a7de-4ce8-a337-3c3f7be95120 for instance with vm_state active and task_state None. [ 1004.602471] env[61970]: DEBUG nova.compute.manager [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1004.602751] env[61970]: DEBUG nova.network.neutron [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1004.611841] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356029, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.612084] env[61970]: WARNING oslo_vmware.common.loopingcall [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] task run outlasted interval by 0.30641700000000005 sec [ 1004.616583] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356027, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073256} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.620569] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1004.622361] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d6afc6-c6cc-411b-a3c6-fcfcaa9037ff {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.631073] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356029, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.650849] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] d38bde37-994b-4d58-b4b1-5f79b59fd2aa/d38bde37-994b-4d58-b4b1-5f79b59fd2aa.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.652917] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-28828239-8708-4857-9c24-088401a05521 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.677509] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 1004.677509] env[61970]: value = "task-1356030" [ 1004.677509] env[61970]: _type = "Task" [ 1004.677509] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.677509] env[61970]: DEBUG nova.policy [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '67a4e38db9894750b16edcf7f188220f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '19dae2316adb4dc980bd972c8c40c034', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 1004.689959] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.079236] env[61970]: DEBUG nova.network.neutron [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Successfully created port: b39ddbde-1199-4f98-9fe2-cd05d19bc96e {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1005.099616] env[61970]: DEBUG nova.compute.manager [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1005.105702] env[61970]: DEBUG oslo_concurrency.lockutils [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.105702] env[61970]: DEBUG oslo_concurrency.lockutils [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.105702] env[61970]: DEBUG nova.network.neutron [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1005.130033] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356029, 'name': ReconfigVM_Task, 'duration_secs': 0.943128} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.130033] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Reconfigured VM instance instance-00000050 to attach disk [datastore2] volume-a78067a3-226b-4aec-bb5b-6fb37d5bfca3/volume-a78067a3-226b-4aec-bb5b-6fb37d5bfca3.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1005.134083] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c96f82e-8b48-4c6e-bcb6-e40a80505f44 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.150332] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1005.150332] env[61970]: value = "task-1356031" [ 1005.150332] env[61970]: _type = "Task" [ 1005.150332] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.154091] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f61302e5-41f0-4916-8e59-e3ed7e24f794 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "00e3287d-24fb-4ea9-b212-54215828d5df" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.154323] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f61302e5-41f0-4916-8e59-e3ed7e24f794 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "00e3287d-24fb-4ea9-b212-54215828d5df" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.154502] env[61970]: DEBUG nova.compute.manager [None req-f61302e5-41f0-4916-8e59-e3ed7e24f794 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1005.155511] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-590c6627-8a12-46fe-ae59-f68794e591f7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.165509] env[61970]: DEBUG nova.compute.manager [None req-f61302e5-41f0-4916-8e59-e3ed7e24f794 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61970) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1005.166447] env[61970]: DEBUG nova.objects.instance [None req-f61302e5-41f0-4916-8e59-e3ed7e24f794 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lazy-loading 'flavor' on Instance uuid 00e3287d-24fb-4ea9-b212-54215828d5df {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.168097] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356031, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.184656] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356030, 'name': ReconfigVM_Task, 'duration_secs': 0.395484} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.184949] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Reconfigured VM instance instance-00000060 to attach disk [datastore2] d38bde37-994b-4d58-b4b1-5f79b59fd2aa/d38bde37-994b-4d58-b4b1-5f79b59fd2aa.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1005.185588] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-85e5507f-a8c0-4475-b4ff-734d46033a30 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.191559] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 1005.191559] env[61970]: value = "task-1356032" [ 1005.191559] env[61970]: _type = "Task" [ 1005.191559] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.201865] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356032, 'name': Rename_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.655178] env[61970]: WARNING nova.network.neutron [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] 8f622fc5-b74f-4582-a6e0-44c90f495750 already exists in list: networks containing: ['8f622fc5-b74f-4582-a6e0-44c90f495750']. ignoring it [ 1005.671621] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356031, 'name': ReconfigVM_Task, 'duration_secs': 0.206791} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.672903] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f61302e5-41f0-4916-8e59-e3ed7e24f794 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1005.676301] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288853', 'volume_id': 'a78067a3-226b-4aec-bb5b-6fb37d5bfca3', 'name': 'volume-a78067a3-226b-4aec-bb5b-6fb37d5bfca3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '7f59000a-94d9-45b6-aa7a-300d95793615', 'attached_at': '', 'detached_at': '', 'volume_id': 'a78067a3-226b-4aec-bb5b-6fb37d5bfca3', 'serial': 'a78067a3-226b-4aec-bb5b-6fb37d5bfca3'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1005.676946] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a45249b0-cbcd-42ab-a0a5-1ea35614f634 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.678792] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b0206b8-6227-4eb5-8a90-4f54094ccf5d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.685848] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1005.685848] env[61970]: value = "task-1356033" [ 1005.685848] env[61970]: _type = "Task" [ 1005.685848] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.688916] env[61970]: DEBUG oslo_vmware.api [None req-f61302e5-41f0-4916-8e59-e3ed7e24f794 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 1005.688916] env[61970]: value = "task-1356034" [ 1005.688916] env[61970]: _type = "Task" [ 1005.688916] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.705400] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356033, 'name': Rename_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.712040] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356032, 'name': Rename_Task, 'duration_secs': 0.140494} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.712287] env[61970]: DEBUG oslo_vmware.api [None req-f61302e5-41f0-4916-8e59-e3ed7e24f794 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356034, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.715257] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1005.715851] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e7f6d2f5-a654-48e5-a22c-c68a636d6d7f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.722289] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 1005.722289] env[61970]: value = "task-1356035" [ 1005.722289] env[61970]: _type = "Task" [ 1005.722289] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.738153] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356035, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.814505] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb81c425-7332-41ce-bbc0-576c637c9604 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.824096] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-253391a7-a1f6-45ea-ba6e-122ee7531bfc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.858304] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c22691-8a9a-4c06-84ae-17794d3593b0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.866885] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10890e69-df9a-4a75-8b2c-1f5a3e62e24d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.884817] env[61970]: DEBUG nova.compute.provider_tree [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.119242] env[61970]: DEBUG nova.compute.manager [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1006.146275] env[61970]: DEBUG nova.virt.hardware [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1006.146630] env[61970]: DEBUG nova.virt.hardware [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1006.146826] env[61970]: DEBUG nova.virt.hardware [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1006.147128] env[61970]: DEBUG nova.virt.hardware [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1006.147387] env[61970]: DEBUG nova.virt.hardware [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1006.147645] env[61970]: DEBUG nova.virt.hardware [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1006.147900] env[61970]: DEBUG nova.virt.hardware [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1006.148163] env[61970]: DEBUG nova.virt.hardware [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1006.148420] env[61970]: DEBUG nova.virt.hardware [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1006.148655] env[61970]: DEBUG nova.virt.hardware [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1006.148941] env[61970]: DEBUG nova.virt.hardware [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1006.150193] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0253c3e5-2a62-4170-9eac-bc203477783c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.154572] env[61970]: DEBUG nova.network.neutron [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Updating instance_info_cache with network_info: [{"id": "4434695d-ddda-4840-b1da-8159c98dcae0", "address": "fa:16:3e:85:27:97", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4434695d-dd", "ovs_interfaceid": "4434695d-ddda-4840-b1da-8159c98dcae0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "90fbe562-a7de-4ce8-a337-3c3f7be95120", "address": "fa:16:3e:91:3a:b6", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90fbe562-a7", "ovs_interfaceid": "90fbe562-a7de-4ce8-a337-3c3f7be95120", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.163414] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2139d20c-3cd8-4b52-9405-63fbb5e8fce4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.198162] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356033, 'name': Rename_Task, 'duration_secs': 0.249693} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.199114] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1006.199578] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c606f903-bec3-4a33-aefb-acca3f908a7e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.204713] env[61970]: DEBUG oslo_vmware.api [None req-f61302e5-41f0-4916-8e59-e3ed7e24f794 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356034, 'name': PowerOffVM_Task, 'duration_secs': 0.255501} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.206332] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f61302e5-41f0-4916-8e59-e3ed7e24f794 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1006.206622] env[61970]: DEBUG nova.compute.manager [None req-f61302e5-41f0-4916-8e59-e3ed7e24f794 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1006.207629] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1006.207629] env[61970]: value = "task-1356036" [ 1006.207629] env[61970]: _type = "Task" [ 1006.207629] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.207850] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6a9560-8178-45f4-a3c4-854768a1b33b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.222113] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356036, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.234820] env[61970]: DEBUG oslo_vmware.api [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356035, 'name': PowerOnVM_Task, 'duration_secs': 0.483326} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.235224] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1006.235539] env[61970]: INFO nova.compute.manager [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Took 11.35 seconds to spawn the instance on the hypervisor. [ 1006.235810] env[61970]: DEBUG nova.compute.manager [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1006.236677] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3b59bfc-fe32-4d21-a5bb-875da841f059 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.388725] env[61970]: DEBUG nova.scheduler.client.report [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1006.573580] env[61970]: DEBUG nova.compute.manager [req-e1fdad52-3067-4a1f-b4f6-ef2fc1c0fad4 req-8880f108-bef4-4176-9258-7e133d3e08f2 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Received event network-changed-90fbe562-a7de-4ce8-a337-3c3f7be95120 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.573868] env[61970]: DEBUG nova.compute.manager [req-e1fdad52-3067-4a1f-b4f6-ef2fc1c0fad4 req-8880f108-bef4-4176-9258-7e133d3e08f2 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Refreshing instance network info cache due to event network-changed-90fbe562-a7de-4ce8-a337-3c3f7be95120. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1006.574201] env[61970]: DEBUG oslo_concurrency.lockutils [req-e1fdad52-3067-4a1f-b4f6-ef2fc1c0fad4 req-8880f108-bef4-4176-9258-7e133d3e08f2 service nova] Acquiring lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.657404] env[61970]: DEBUG oslo_concurrency.lockutils [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.658391] env[61970]: DEBUG oslo_concurrency.lockutils [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.658601] env[61970]: DEBUG oslo_concurrency.lockutils [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.659048] env[61970]: DEBUG oslo_concurrency.lockutils [req-e1fdad52-3067-4a1f-b4f6-ef2fc1c0fad4 req-8880f108-bef4-4176-9258-7e133d3e08f2 service nova] Acquired lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.659306] env[61970]: DEBUG nova.network.neutron [req-e1fdad52-3067-4a1f-b4f6-ef2fc1c0fad4 req-8880f108-bef4-4176-9258-7e133d3e08f2 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Refreshing network info cache for port 90fbe562-a7de-4ce8-a337-3c3f7be95120 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1006.664022] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0a1829a-6a2a-413b-8269-0e95339fea35 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.681331] env[61970]: DEBUG nova.virt.hardware [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1006.681665] env[61970]: DEBUG nova.virt.hardware [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1006.681880] env[61970]: DEBUG nova.virt.hardware [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1006.682334] env[61970]: DEBUG nova.virt.hardware [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1006.682334] env[61970]: DEBUG nova.virt.hardware [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1006.682449] env[61970]: DEBUG nova.virt.hardware [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1006.682691] env[61970]: DEBUG nova.virt.hardware [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1006.682878] env[61970]: DEBUG nova.virt.hardware [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1006.683078] env[61970]: DEBUG nova.virt.hardware [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1006.683255] env[61970]: DEBUG nova.virt.hardware [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1006.683470] env[61970]: DEBUG nova.virt.hardware [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1006.689896] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Reconfiguring VM to attach interface {{(pid=61970) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1006.690868] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-668804e2-3668-4295-8203-00ee56a0b509 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.713028] env[61970]: DEBUG oslo_vmware.api [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 1006.713028] env[61970]: value = "task-1356037" [ 1006.713028] env[61970]: _type = "Task" [ 1006.713028] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.728898] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356036, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.730575] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f61302e5-41f0-4916-8e59-e3ed7e24f794 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "00e3287d-24fb-4ea9-b212-54215828d5df" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.576s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.734020] env[61970]: DEBUG oslo_vmware.api [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356037, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.734020] env[61970]: DEBUG nova.network.neutron [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Successfully updated port: b39ddbde-1199-4f98-9fe2-cd05d19bc96e {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1006.755210] env[61970]: INFO nova.compute.manager [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Took 20.82 seconds to build instance. [ 1006.893673] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.312s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.894316] env[61970]: DEBUG nova.compute.manager [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1007.229469] env[61970]: DEBUG oslo_vmware.api [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356037, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.229804] env[61970]: DEBUG oslo_vmware.api [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356036, 'name': PowerOnVM_Task, 'duration_secs': 0.760402} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.230122] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1007.234692] env[61970]: DEBUG oslo_concurrency.lockutils [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "refresh_cache-31f9d37b-f4ff-4a8b-9477-8253f8305020" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.234896] env[61970]: DEBUG oslo_concurrency.lockutils [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquired lock "refresh_cache-31f9d37b-f4ff-4a8b-9477-8253f8305020" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.235141] env[61970]: DEBUG nova.network.neutron [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1007.258245] env[61970]: DEBUG oslo_concurrency.lockutils [None req-44436d46-1893-4830-9d12-f29727408678 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "d38bde37-994b-4d58-b4b1-5f79b59fd2aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.330s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.377359] env[61970]: DEBUG nova.compute.manager [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1007.378573] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f011cf-98a8-4d12-9f2e-94e339dbfa77 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.399501] env[61970]: DEBUG nova.compute.utils [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1007.400856] env[61970]: DEBUG nova.compute.manager [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1007.401439] env[61970]: DEBUG nova.network.neutron [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1007.482050] env[61970]: DEBUG nova.policy [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'abe55db615be47e09602ec51db877533', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '16e2af65733d46a1a3a2bb5613b3d63e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 1007.546439] env[61970]: DEBUG nova.network.neutron [req-e1fdad52-3067-4a1f-b4f6-ef2fc1c0fad4 req-8880f108-bef4-4176-9258-7e133d3e08f2 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Updated VIF entry in instance network info cache for port 90fbe562-a7de-4ce8-a337-3c3f7be95120. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1007.547198] env[61970]: DEBUG nova.network.neutron [req-e1fdad52-3067-4a1f-b4f6-ef2fc1c0fad4 req-8880f108-bef4-4176-9258-7e133d3e08f2 service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Updating instance_info_cache with network_info: [{"id": "4434695d-ddda-4840-b1da-8159c98dcae0", "address": "fa:16:3e:85:27:97", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4434695d-dd", "ovs_interfaceid": "4434695d-ddda-4840-b1da-8159c98dcae0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "90fbe562-a7de-4ce8-a337-3c3f7be95120", "address": "fa:16:3e:91:3a:b6", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90fbe562-a7", "ovs_interfaceid": "90fbe562-a7de-4ce8-a337-3c3f7be95120", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.720891] env[61970]: DEBUG oslo_vmware.api [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356037, 'name': ReconfigVM_Task, 'duration_secs': 0.612443} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.722693] env[61970]: DEBUG oslo_concurrency.lockutils [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.723542] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Reconfigured VM to attach interface {{(pid=61970) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1007.803329] env[61970]: DEBUG nova.network.neutron [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1007.879771] env[61970]: DEBUG nova.compute.manager [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Stashing vm_state: active {{(pid=61970) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1007.896904] env[61970]: DEBUG oslo_concurrency.lockutils [None req-06201b0d-5ed5-431f-9ae4-66c0715e1d27 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "7f59000a-94d9-45b6-aa7a-300d95793615" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 32.007s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.904045] env[61970]: DEBUG nova.compute.manager [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1007.907413] env[61970]: DEBUG nova.network.neutron [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Successfully created port: babd8ae5-2695-4eb8-97d1-e9db1f6c518e {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1008.049892] env[61970]: DEBUG oslo_concurrency.lockutils [req-e1fdad52-3067-4a1f-b4f6-ef2fc1c0fad4 req-8880f108-bef4-4176-9258-7e133d3e08f2 service nova] Releasing lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.083715] env[61970]: DEBUG nova.network.neutron [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Updating instance_info_cache with network_info: [{"id": "b39ddbde-1199-4f98-9fe2-cd05d19bc96e", "address": "fa:16:3e:be:c5:78", "network": {"id": "42def9ae-4eed-41e5-a538-81eaced2cdd4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2047751481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19dae2316adb4dc980bd972c8c40c034", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c1b8689-a9b4-4972-beb9-6a1c8de1dc88", "external-id": "nsx-vlan-transportzone-455", "segmentation_id": 455, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb39ddbde-11", "ovs_interfaceid": "b39ddbde-1199-4f98-9fe2-cd05d19bc96e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.229306] env[61970]: DEBUG oslo_concurrency.lockutils [None req-23cb2dbd-7990-4e5d-be75-af5ac360f64f tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "interface-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8-90fbe562-a7de-4ce8-a337-3c3f7be95120" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.217s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.230576] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "00e3287d-24fb-4ea9-b212-54215828d5df" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.231389] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "00e3287d-24fb-4ea9-b212-54215828d5df" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.231641] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "00e3287d-24fb-4ea9-b212-54215828d5df-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.231857] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "00e3287d-24fb-4ea9-b212-54215828d5df-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.232036] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "00e3287d-24fb-4ea9-b212-54215828d5df-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.234192] env[61970]: INFO nova.compute.manager [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Terminating instance [ 1008.235964] env[61970]: DEBUG nova.compute.manager [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1008.236192] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1008.237115] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe00f837-c26d-43f7-b7b0-7d66bcc54d30 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.245403] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1008.245718] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e5d92e8-9a72-40ce-8e7d-51d4db6e23cc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.404129] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.404129] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.586172] env[61970]: DEBUG oslo_concurrency.lockutils [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Releasing lock "refresh_cache-31f9d37b-f4ff-4a8b-9477-8253f8305020" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.586521] env[61970]: DEBUG nova.compute.manager [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Instance network_info: |[{"id": "b39ddbde-1199-4f98-9fe2-cd05d19bc96e", "address": "fa:16:3e:be:c5:78", "network": {"id": "42def9ae-4eed-41e5-a538-81eaced2cdd4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2047751481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19dae2316adb4dc980bd972c8c40c034", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c1b8689-a9b4-4972-beb9-6a1c8de1dc88", "external-id": "nsx-vlan-transportzone-455", "segmentation_id": 455, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb39ddbde-11", "ovs_interfaceid": "b39ddbde-1199-4f98-9fe2-cd05d19bc96e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1008.587050] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:be:c5:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9c1b8689-a9b4-4972-beb9-6a1c8de1dc88', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b39ddbde-1199-4f98-9fe2-cd05d19bc96e', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1008.595022] env[61970]: DEBUG oslo.service.loopingcall [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1008.595847] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1008.596253] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-486dd694-193c-4bfe-8d19-1daf258ca76a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.619057] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1008.619057] env[61970]: value = "task-1356039" [ 1008.619057] env[61970]: _type = "Task" [ 1008.619057] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.627084] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356039, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.719208] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1008.719512] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1008.719740] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleting the datastore file [datastore2] 00e3287d-24fb-4ea9-b212-54215828d5df {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1008.720025] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-346adbff-487f-4401-9d49-6d09a96a6ba1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.727118] env[61970]: DEBUG oslo_vmware.api [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for the task: (returnval){ [ 1008.727118] env[61970]: value = "task-1356040" [ 1008.727118] env[61970]: _type = "Task" [ 1008.727118] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.736561] env[61970]: DEBUG oslo_vmware.api [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356040, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.749107] env[61970]: DEBUG nova.compute.manager [req-340475ff-dabd-4c14-85ec-18f57a243c06 req-dbb346ff-46e1-4f01-ad10-88f56af97612 service nova] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Received event network-vif-plugged-b39ddbde-1199-4f98-9fe2-cd05d19bc96e {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.749322] env[61970]: DEBUG oslo_concurrency.lockutils [req-340475ff-dabd-4c14-85ec-18f57a243c06 req-dbb346ff-46e1-4f01-ad10-88f56af97612 service nova] Acquiring lock "31f9d37b-f4ff-4a8b-9477-8253f8305020-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.749528] env[61970]: DEBUG oslo_concurrency.lockutils [req-340475ff-dabd-4c14-85ec-18f57a243c06 req-dbb346ff-46e1-4f01-ad10-88f56af97612 service nova] Lock "31f9d37b-f4ff-4a8b-9477-8253f8305020-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.749704] env[61970]: DEBUG oslo_concurrency.lockutils [req-340475ff-dabd-4c14-85ec-18f57a243c06 req-dbb346ff-46e1-4f01-ad10-88f56af97612 service nova] Lock "31f9d37b-f4ff-4a8b-9477-8253f8305020-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.749878] env[61970]: DEBUG nova.compute.manager [req-340475ff-dabd-4c14-85ec-18f57a243c06 req-dbb346ff-46e1-4f01-ad10-88f56af97612 service nova] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] No waiting events found dispatching network-vif-plugged-b39ddbde-1199-4f98-9fe2-cd05d19bc96e {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1008.750156] env[61970]: WARNING nova.compute.manager [req-340475ff-dabd-4c14-85ec-18f57a243c06 req-dbb346ff-46e1-4f01-ad10-88f56af97612 service nova] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Received unexpected event network-vif-plugged-b39ddbde-1199-4f98-9fe2-cd05d19bc96e for instance with vm_state building and task_state spawning. [ 1008.750331] env[61970]: DEBUG nova.compute.manager [req-340475ff-dabd-4c14-85ec-18f57a243c06 req-dbb346ff-46e1-4f01-ad10-88f56af97612 service nova] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Received event network-changed-b39ddbde-1199-4f98-9fe2-cd05d19bc96e {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.750491] env[61970]: DEBUG nova.compute.manager [req-340475ff-dabd-4c14-85ec-18f57a243c06 req-dbb346ff-46e1-4f01-ad10-88f56af97612 service nova] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Refreshing instance network info cache due to event network-changed-b39ddbde-1199-4f98-9fe2-cd05d19bc96e. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1008.750683] env[61970]: DEBUG oslo_concurrency.lockutils [req-340475ff-dabd-4c14-85ec-18f57a243c06 req-dbb346ff-46e1-4f01-ad10-88f56af97612 service nova] Acquiring lock "refresh_cache-31f9d37b-f4ff-4a8b-9477-8253f8305020" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.750820] env[61970]: DEBUG oslo_concurrency.lockutils [req-340475ff-dabd-4c14-85ec-18f57a243c06 req-dbb346ff-46e1-4f01-ad10-88f56af97612 service nova] Acquired lock "refresh_cache-31f9d37b-f4ff-4a8b-9477-8253f8305020" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.750979] env[61970]: DEBUG nova.network.neutron [req-340475ff-dabd-4c14-85ec-18f57a243c06 req-dbb346ff-46e1-4f01-ad10-88f56af97612 service nova] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Refreshing network info cache for port b39ddbde-1199-4f98-9fe2-cd05d19bc96e {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1008.909463] env[61970]: INFO nova.compute.claims [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1008.914681] env[61970]: DEBUG nova.compute.manager [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1008.942400] env[61970]: DEBUG nova.virt.hardware [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1008.942666] env[61970]: DEBUG nova.virt.hardware [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1008.942807] env[61970]: DEBUG nova.virt.hardware [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1008.942999] env[61970]: DEBUG nova.virt.hardware [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1008.943151] env[61970]: DEBUG nova.virt.hardware [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1008.943300] env[61970]: DEBUG nova.virt.hardware [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1008.943512] env[61970]: DEBUG nova.virt.hardware [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1008.943677] env[61970]: DEBUG nova.virt.hardware [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1008.943849] env[61970]: DEBUG nova.virt.hardware [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1008.944043] env[61970]: DEBUG nova.virt.hardware [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1008.944203] env[61970]: DEBUG nova.virt.hardware [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1008.945381] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6762e5-f17b-45f8-a50d-d0718131bfcc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.953450] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84738d0c-da96-46d8-b27d-4686670454ec {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.128954] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356039, 'name': CreateVM_Task, 'duration_secs': 0.464363} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.129891] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1009.129891] env[61970]: DEBUG oslo_concurrency.lockutils [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.130061] env[61970]: DEBUG oslo_concurrency.lockutils [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.130391] env[61970]: DEBUG oslo_concurrency.lockutils [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1009.130657] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45294fab-5362-4ff3-a4f4-49ceb026d0e5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.135552] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1009.135552] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52bf2d9b-e47c-d136-362b-7c5a40a55efe" [ 1009.135552] env[61970]: _type = "Task" [ 1009.135552] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.143626] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52bf2d9b-e47c-d136-362b-7c5a40a55efe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.237817] env[61970]: DEBUG oslo_vmware.api [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Task: {'id': task-1356040, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190949} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.238059] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1009.238258] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1009.238437] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1009.238610] env[61970]: INFO nova.compute.manager [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Took 1.00 seconds to destroy the instance on the hypervisor. [ 1009.238864] env[61970]: DEBUG oslo.service.loopingcall [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.239404] env[61970]: DEBUG nova.compute.manager [-] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1009.239515] env[61970]: DEBUG nova.network.neutron [-] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1009.420195] env[61970]: INFO nova.compute.resource_tracker [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Updating resource usage from migration 63900b78-7846-41b5-89aa-ce6edab35b4e [ 1009.463264] env[61970]: DEBUG nova.network.neutron [req-340475ff-dabd-4c14-85ec-18f57a243c06 req-dbb346ff-46e1-4f01-ad10-88f56af97612 service nova] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Updated VIF entry in instance network info cache for port b39ddbde-1199-4f98-9fe2-cd05d19bc96e. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1009.463829] env[61970]: DEBUG nova.network.neutron [req-340475ff-dabd-4c14-85ec-18f57a243c06 req-dbb346ff-46e1-4f01-ad10-88f56af97612 service nova] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Updating instance_info_cache with network_info: [{"id": "b39ddbde-1199-4f98-9fe2-cd05d19bc96e", "address": "fa:16:3e:be:c5:78", "network": {"id": "42def9ae-4eed-41e5-a538-81eaced2cdd4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2047751481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19dae2316adb4dc980bd972c8c40c034", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c1b8689-a9b4-4972-beb9-6a1c8de1dc88", "external-id": "nsx-vlan-transportzone-455", "segmentation_id": 455, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb39ddbde-11", "ovs_interfaceid": "b39ddbde-1199-4f98-9fe2-cd05d19bc96e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.624128] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af8c8425-2fc7-4a90-be67-f6d40bc9c7bd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.632980] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a340f1-03e5-41d8-9bea-90fca3caed65 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.644371] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52bf2d9b-e47c-d136-362b-7c5a40a55efe, 'name': SearchDatastore_Task, 'duration_secs': 0.028015} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.669861] env[61970]: DEBUG oslo_concurrency.lockutils [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.670153] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1009.670401] env[61970]: DEBUG oslo_concurrency.lockutils [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.670563] env[61970]: DEBUG oslo_concurrency.lockutils [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.670756] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1009.671602] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4301e35-9f67-48c8-9b62-ebf92b3fe59f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.674043] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c24a6d7-ebc6-4759-9f19-43c35bc62fc4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.682143] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69272cd2-756c-4f0b-a9bc-970ec343fd32 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.687017] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1009.687684] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1009.688311] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68b5e67e-44a9-4b4a-8e03-82a80a675965 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.698389] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "interface-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8-90fbe562-a7de-4ce8-a337-3c3f7be95120" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.698643] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "interface-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8-90fbe562-a7de-4ce8-a337-3c3f7be95120" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.700158] env[61970]: DEBUG nova.compute.provider_tree [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.704952] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1009.704952] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d4f4a2-e377-4464-ff57-3a6a8b867c4e" [ 1009.704952] env[61970]: _type = "Task" [ 1009.704952] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.714409] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d4f4a2-e377-4464-ff57-3a6a8b867c4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.722063] env[61970]: DEBUG nova.network.neutron [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Successfully updated port: babd8ae5-2695-4eb8-97d1-e9db1f6c518e {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1009.966544] env[61970]: DEBUG oslo_concurrency.lockutils [req-340475ff-dabd-4c14-85ec-18f57a243c06 req-dbb346ff-46e1-4f01-ad10-88f56af97612 service nova] Releasing lock "refresh_cache-31f9d37b-f4ff-4a8b-9477-8253f8305020" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.110388] env[61970]: DEBUG nova.network.neutron [-] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.202722] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.203026] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.204039] env[61970]: DEBUG nova.scheduler.client.report [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1010.208440] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc6bc486-da4e-4696-97e6-545f87a9fc1d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.222507] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d4f4a2-e377-4464-ff57-3a6a8b867c4e, 'name': SearchDatastore_Task, 'duration_secs': 0.026239} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.236834] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Acquiring lock "refresh_cache-19c0bfcb-1964-4719-98b1-d9334751ace2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.236980] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Acquired lock "refresh_cache-19c0bfcb-1964-4719-98b1-d9334751ace2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.237144] env[61970]: DEBUG nova.network.neutron [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1010.238848] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b996e60-38b4-42e1-89b2-d86bfa9434ff {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.242465] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a4546c-7a13-4390-ae75-985def403dfd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.249413] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1010.249413] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525488c6-5b30-23d1-6ad8-7bc34b1a2f64" [ 1010.249413] env[61970]: _type = "Task" [ 1010.249413] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.271723] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Reconfiguring VM to detach interface {{(pid=61970) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1010.275421] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f07ec7b-89d3-484c-855f-8748e54c4749 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.299126] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525488c6-5b30-23d1-6ad8-7bc34b1a2f64, 'name': SearchDatastore_Task, 'duration_secs': 0.022139} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.300583] env[61970]: DEBUG oslo_concurrency.lockutils [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.300864] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 31f9d37b-f4ff-4a8b-9477-8253f8305020/31f9d37b-f4ff-4a8b-9477-8253f8305020.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1010.301254] env[61970]: DEBUG oslo_vmware.api [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 1010.301254] env[61970]: value = "task-1356041" [ 1010.301254] env[61970]: _type = "Task" [ 1010.301254] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.301489] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-34f8e744-a581-454b-bf25-572c6c1d9a73 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.314361] env[61970]: DEBUG oslo_vmware.api [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356041, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.315895] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1010.315895] env[61970]: value = "task-1356042" [ 1010.315895] env[61970]: _type = "Task" [ 1010.315895] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.324523] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356042, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.613920] env[61970]: INFO nova.compute.manager [-] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Took 1.37 seconds to deallocate network for instance. [ 1010.713697] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.310s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.714024] env[61970]: INFO nova.compute.manager [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Migrating [ 1010.770639] env[61970]: DEBUG nova.network.neutron [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1010.780500] env[61970]: DEBUG nova.compute.manager [req-322f87bd-d1d4-45bf-8416-74a5d13b017c req-bf9b1714-9169-4134-8c03-c4c2a758ba59 service nova] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Received event network-vif-plugged-babd8ae5-2695-4eb8-97d1-e9db1f6c518e {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1010.780747] env[61970]: DEBUG oslo_concurrency.lockutils [req-322f87bd-d1d4-45bf-8416-74a5d13b017c req-bf9b1714-9169-4134-8c03-c4c2a758ba59 service nova] Acquiring lock "19c0bfcb-1964-4719-98b1-d9334751ace2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.780997] env[61970]: DEBUG oslo_concurrency.lockutils [req-322f87bd-d1d4-45bf-8416-74a5d13b017c req-bf9b1714-9169-4134-8c03-c4c2a758ba59 service nova] Lock "19c0bfcb-1964-4719-98b1-d9334751ace2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.781202] env[61970]: DEBUG oslo_concurrency.lockutils [req-322f87bd-d1d4-45bf-8416-74a5d13b017c req-bf9b1714-9169-4134-8c03-c4c2a758ba59 service nova] Lock "19c0bfcb-1964-4719-98b1-d9334751ace2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.781377] env[61970]: DEBUG nova.compute.manager [req-322f87bd-d1d4-45bf-8416-74a5d13b017c req-bf9b1714-9169-4134-8c03-c4c2a758ba59 service nova] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] No waiting events found dispatching network-vif-plugged-babd8ae5-2695-4eb8-97d1-e9db1f6c518e {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1010.781546] env[61970]: WARNING nova.compute.manager [req-322f87bd-d1d4-45bf-8416-74a5d13b017c req-bf9b1714-9169-4134-8c03-c4c2a758ba59 service nova] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Received unexpected event network-vif-plugged-babd8ae5-2695-4eb8-97d1-e9db1f6c518e for instance with vm_state building and task_state spawning. [ 1010.781773] env[61970]: DEBUG nova.compute.manager [req-322f87bd-d1d4-45bf-8416-74a5d13b017c req-bf9b1714-9169-4134-8c03-c4c2a758ba59 service nova] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Received event network-vif-deleted-f7943b8e-3fab-4cfc-b59d-5a7a01efb9ba {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1010.781967] env[61970]: DEBUG nova.compute.manager [req-322f87bd-d1d4-45bf-8416-74a5d13b017c req-bf9b1714-9169-4134-8c03-c4c2a758ba59 service nova] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Received event network-changed-babd8ae5-2695-4eb8-97d1-e9db1f6c518e {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1010.782158] env[61970]: DEBUG nova.compute.manager [req-322f87bd-d1d4-45bf-8416-74a5d13b017c req-bf9b1714-9169-4134-8c03-c4c2a758ba59 service nova] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Refreshing instance network info cache due to event network-changed-babd8ae5-2695-4eb8-97d1-e9db1f6c518e. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1010.782342] env[61970]: DEBUG oslo_concurrency.lockutils [req-322f87bd-d1d4-45bf-8416-74a5d13b017c req-bf9b1714-9169-4134-8c03-c4c2a758ba59 service nova] Acquiring lock "refresh_cache-19c0bfcb-1964-4719-98b1-d9334751ace2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.813858] env[61970]: DEBUG oslo_vmware.api [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356041, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.824257] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356042, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.946139] env[61970]: DEBUG nova.network.neutron [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Updating instance_info_cache with network_info: [{"id": "babd8ae5-2695-4eb8-97d1-e9db1f6c518e", "address": "fa:16:3e:70:d7:9c", "network": {"id": "a72082ae-fd5a-4947-971c-4fbb4bfe3beb", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1764952967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16e2af65733d46a1a3a2bb5613b3d63e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbabd8ae5-26", "ovs_interfaceid": "babd8ae5-2695-4eb8-97d1-e9db1f6c518e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.121095] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.121575] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.121909] env[61970]: DEBUG nova.objects.instance [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lazy-loading 'resources' on Instance uuid 00e3287d-24fb-4ea9-b212-54215828d5df {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.229031] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "refresh_cache-d38bde37-994b-4d58-b4b1-5f79b59fd2aa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.229295] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "refresh_cache-d38bde37-994b-4d58-b4b1-5f79b59fd2aa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.229422] env[61970]: DEBUG nova.network.neutron [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1011.314439] env[61970]: DEBUG oslo_vmware.api [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356041, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.325098] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356042, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.907578} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.325398] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 31f9d37b-f4ff-4a8b-9477-8253f8305020/31f9d37b-f4ff-4a8b-9477-8253f8305020.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1011.325638] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1011.325923] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9be76cb5-96a4-45e4-be3c-b648db48cb03 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.331963] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1011.331963] env[61970]: value = "task-1356043" [ 1011.331963] env[61970]: _type = "Task" [ 1011.331963] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.341502] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356043, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.449876] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Releasing lock "refresh_cache-19c0bfcb-1964-4719-98b1-d9334751ace2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.450305] env[61970]: DEBUG nova.compute.manager [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Instance network_info: |[{"id": "babd8ae5-2695-4eb8-97d1-e9db1f6c518e", "address": "fa:16:3e:70:d7:9c", "network": {"id": "a72082ae-fd5a-4947-971c-4fbb4bfe3beb", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1764952967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16e2af65733d46a1a3a2bb5613b3d63e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbabd8ae5-26", "ovs_interfaceid": "babd8ae5-2695-4eb8-97d1-e9db1f6c518e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1011.450684] env[61970]: DEBUG oslo_concurrency.lockutils [req-322f87bd-d1d4-45bf-8416-74a5d13b017c req-bf9b1714-9169-4134-8c03-c4c2a758ba59 service nova] Acquired lock "refresh_cache-19c0bfcb-1964-4719-98b1-d9334751ace2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.450830] env[61970]: DEBUG nova.network.neutron [req-322f87bd-d1d4-45bf-8416-74a5d13b017c req-bf9b1714-9169-4134-8c03-c4c2a758ba59 service nova] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Refreshing network info cache for port babd8ae5-2695-4eb8-97d1-e9db1f6c518e {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1011.453622] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:d7:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '051f343d-ac4f-4070-a26d-467603122c81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'babd8ae5-2695-4eb8-97d1-e9db1f6c518e', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1011.459876] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Creating folder: Project (16e2af65733d46a1a3a2bb5613b3d63e). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1011.462993] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0b263810-f1e9-4045-871c-f1e12516fc9a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.474850] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Created folder: Project (16e2af65733d46a1a3a2bb5613b3d63e) in parent group-v288740. [ 1011.474994] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Creating folder: Instances. Parent ref: group-v288864. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1011.475233] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2f71234c-ea24-47f2-a102-50852bf08998 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.485259] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Created folder: Instances in parent group-v288864. [ 1011.485499] env[61970]: DEBUG oslo.service.loopingcall [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.485688] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1011.485890] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0a749fa7-eea1-43d9-b453-4bd1f9f6ad1c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.507274] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1011.507274] env[61970]: value = "task-1356046" [ 1011.507274] env[61970]: _type = "Task" [ 1011.507274] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.516778] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356046, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.723098] env[61970]: DEBUG nova.network.neutron [req-322f87bd-d1d4-45bf-8416-74a5d13b017c req-bf9b1714-9169-4134-8c03-c4c2a758ba59 service nova] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Updated VIF entry in instance network info cache for port babd8ae5-2695-4eb8-97d1-e9db1f6c518e. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1011.723494] env[61970]: DEBUG nova.network.neutron [req-322f87bd-d1d4-45bf-8416-74a5d13b017c req-bf9b1714-9169-4134-8c03-c4c2a758ba59 service nova] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Updating instance_info_cache with network_info: [{"id": "babd8ae5-2695-4eb8-97d1-e9db1f6c518e", "address": "fa:16:3e:70:d7:9c", "network": {"id": "a72082ae-fd5a-4947-971c-4fbb4bfe3beb", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1764952967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16e2af65733d46a1a3a2bb5613b3d63e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbabd8ae5-26", "ovs_interfaceid": "babd8ae5-2695-4eb8-97d1-e9db1f6c518e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.788841] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e04976c-d9fa-42bb-89e0-03501c34b098 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.798582] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e8982a8-8bc8-42cd-bcbc-5d248413e627 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.834153] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd6d853-efec-4d74-8ebe-9917674b242b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.843672] env[61970]: DEBUG oslo_vmware.api [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356041, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.851304] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c923a9d8-c58d-41d1-bb97-509535ab2c4c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.856762] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356043, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.14302} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.857056] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1011.858179] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f157973b-aa17-4b6d-bc64-4fc6268ecca7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.868294] env[61970]: DEBUG nova.compute.provider_tree [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.891082] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 31f9d37b-f4ff-4a8b-9477-8253f8305020/31f9d37b-f4ff-4a8b-9477-8253f8305020.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1011.892056] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63805db6-573b-4d4b-bbc8-119eb4b2fc36 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.914241] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1011.914241] env[61970]: value = "task-1356047" [ 1011.914241] env[61970]: _type = "Task" [ 1011.914241] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.922486] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356047, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.979245] env[61970]: DEBUG nova.network.neutron [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Updating instance_info_cache with network_info: [{"id": "62d22328-b863-41dc-816f-1ce3f7a53eed", "address": "fa:16:3e:53:70:05", "network": {"id": "62edd125-b77c-4418-a465-1dcb3a0733c1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-371307366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6a4997482a4834a30c79c4cbab2cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62d22328-b8", "ovs_interfaceid": "62d22328-b863-41dc-816f-1ce3f7a53eed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.018722] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356046, 'name': CreateVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.228537] env[61970]: DEBUG oslo_concurrency.lockutils [req-322f87bd-d1d4-45bf-8416-74a5d13b017c req-bf9b1714-9169-4134-8c03-c4c2a758ba59 service nova] Releasing lock "refresh_cache-19c0bfcb-1964-4719-98b1-d9334751ace2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.315095] env[61970]: DEBUG oslo_vmware.api [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356041, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.374585] env[61970]: DEBUG nova.scheduler.client.report [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.426122] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356047, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.482248] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "refresh_cache-d38bde37-994b-4d58-b4b1-5f79b59fd2aa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.518032] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356046, 'name': CreateVM_Task, 'duration_secs': 0.586519} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.518233] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1012.518875] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.519062] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.519391] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1012.519660] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4933506-25a5-4f8e-b73f-b25239dec06a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.524418] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Waiting for the task: (returnval){ [ 1012.524418] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52626d64-7d0f-1105-0e5a-402ad9055550" [ 1012.524418] env[61970]: _type = "Task" [ 1012.524418] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.532593] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52626d64-7d0f-1105-0e5a-402ad9055550, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.816065] env[61970]: DEBUG oslo_vmware.api [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356041, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.879764] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.758s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.898080] env[61970]: INFO nova.scheduler.client.report [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Deleted allocations for instance 00e3287d-24fb-4ea9-b212-54215828d5df [ 1012.924908] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356047, 'name': ReconfigVM_Task, 'duration_secs': 0.665647} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.925402] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 31f9d37b-f4ff-4a8b-9477-8253f8305020/31f9d37b-f4ff-4a8b-9477-8253f8305020.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1012.926088] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d4fdbee1-5694-405f-981c-75db05e5207d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.931993] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1012.931993] env[61970]: value = "task-1356048" [ 1012.931993] env[61970]: _type = "Task" [ 1012.931993] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.941946] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356048, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.034871] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52626d64-7d0f-1105-0e5a-402ad9055550, 'name': SearchDatastore_Task, 'duration_secs': 0.014744} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.035192] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.035435] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1013.035663] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.035811] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.035990] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1013.036283] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dad52792-e805-42ed-b99f-fd597f19a17d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.044603] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1013.044777] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1013.045519] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-872a2304-95f5-4cf8-b6d8-e7ab83951261 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.050459] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Waiting for the task: (returnval){ [ 1013.050459] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5215b3b9-c7ea-f1ad-2663-9ce0fba5568e" [ 1013.050459] env[61970]: _type = "Task" [ 1013.050459] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.058998] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5215b3b9-c7ea-f1ad-2663-9ce0fba5568e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.317079] env[61970]: DEBUG oslo_vmware.api [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356041, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.405934] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6370c23a-b665-406e-827b-182ee6c47c99 tempest-ServersTestJSON-1739972093 tempest-ServersTestJSON-1739972093-project-member] Lock "00e3287d-24fb-4ea9-b212-54215828d5df" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.174s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.442834] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356048, 'name': Rename_Task, 'duration_secs': 0.227025} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.443172] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1013.443433] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-185edc83-4061-4cb0-a26d-586c789bf039 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.449789] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1013.449789] env[61970]: value = "task-1356049" [ 1013.449789] env[61970]: _type = "Task" [ 1013.449789] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.458701] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356049, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.562727] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5215b3b9-c7ea-f1ad-2663-9ce0fba5568e, 'name': SearchDatastore_Task, 'duration_secs': 0.017299} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.563761] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f650f78-4e07-43fa-bd3b-d13e5888d972 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.569718] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Waiting for the task: (returnval){ [ 1013.569718] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52690973-42a9-1b30-a75d-7f3144d76c40" [ 1013.569718] env[61970]: _type = "Task" [ 1013.569718] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.579068] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52690973-42a9-1b30-a75d-7f3144d76c40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.817833] env[61970]: DEBUG oslo_vmware.api [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356041, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.962052] env[61970]: DEBUG oslo_vmware.api [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356049, 'name': PowerOnVM_Task, 'duration_secs': 0.511449} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.962443] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1013.962702] env[61970]: INFO nova.compute.manager [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Took 7.84 seconds to spawn the instance on the hypervisor. [ 1013.962944] env[61970]: DEBUG nova.compute.manager [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1013.963918] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be2bb1ae-f77b-4dcd-ae5d-808dfdfba338 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.998177] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8453cb17-3209-45e1-8851-b1e8fe6511b8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.020178] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Updating instance 'd38bde37-994b-4d58-b4b1-5f79b59fd2aa' progress to 0 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1014.082396] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52690973-42a9-1b30-a75d-7f3144d76c40, 'name': SearchDatastore_Task, 'duration_secs': 0.013588} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.082691] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.082964] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 19c0bfcb-1964-4719-98b1-d9334751ace2/19c0bfcb-1964-4719-98b1-d9334751ace2.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1014.083260] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-adf6d3fd-1fcb-4a4d-9801-8751bdbd648d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.090108] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Waiting for the task: (returnval){ [ 1014.090108] env[61970]: value = "task-1356050" [ 1014.090108] env[61970]: _type = "Task" [ 1014.090108] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.099232] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': task-1356050, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.130593] env[61970]: DEBUG oslo_vmware.rw_handles [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52552c28-61b0-83ac-075d-879f4acc9ebc/disk-0.vmdk. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1014.131330] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2556d3da-0563-49a7-b36e-5a2837f59a4e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.137667] env[61970]: DEBUG oslo_vmware.rw_handles [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52552c28-61b0-83ac-075d-879f4acc9ebc/disk-0.vmdk is in state: ready. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1014.137827] env[61970]: ERROR oslo_vmware.rw_handles [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52552c28-61b0-83ac-075d-879f4acc9ebc/disk-0.vmdk due to incomplete transfer. [ 1014.138066] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-87a0f9b3-54f6-4af3-9f96-8a7e5b23f72c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.146141] env[61970]: DEBUG oslo_vmware.rw_handles [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52552c28-61b0-83ac-075d-879f4acc9ebc/disk-0.vmdk. {{(pid=61970) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1014.147025] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Uploaded image d497d9c2-91aa-45aa-90d9-3927851fe30e to the Glance image server {{(pid=61970) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1014.148970] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Destroying the VM {{(pid=61970) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1014.149554] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-fd270a85-a9ad-4ca3-94c5-653c137fd549 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.155401] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1014.155401] env[61970]: value = "task-1356051" [ 1014.155401] env[61970]: _type = "Task" [ 1014.155401] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.164320] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356051, 'name': Destroy_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.319592] env[61970]: DEBUG oslo_vmware.api [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356041, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.482282] env[61970]: INFO nova.compute.manager [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Took 13.19 seconds to build instance. [ 1014.528752] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1014.529070] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7815b836-bd1b-4a43-8d5c-0241669a801a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.538096] env[61970]: DEBUG oslo_vmware.api [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 1014.538096] env[61970]: value = "task-1356052" [ 1014.538096] env[61970]: _type = "Task" [ 1014.538096] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.547500] env[61970]: DEBUG oslo_vmware.api [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356052, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.600026] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': task-1356050, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.666730] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356051, 'name': Destroy_Task} progress is 33%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.819213] env[61970]: DEBUG oslo_vmware.api [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356041, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.866603] env[61970]: DEBUG nova.compute.manager [req-55eebb6b-935e-401d-b833-1b5d601f62de req-acbbd6cb-1a40-4c53-9595-c876c4bcd2ff service nova] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Received event network-changed-b39ddbde-1199-4f98-9fe2-cd05d19bc96e {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.866871] env[61970]: DEBUG nova.compute.manager [req-55eebb6b-935e-401d-b833-1b5d601f62de req-acbbd6cb-1a40-4c53-9595-c876c4bcd2ff service nova] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Refreshing instance network info cache due to event network-changed-b39ddbde-1199-4f98-9fe2-cd05d19bc96e. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1014.867120] env[61970]: DEBUG oslo_concurrency.lockutils [req-55eebb6b-935e-401d-b833-1b5d601f62de req-acbbd6cb-1a40-4c53-9595-c876c4bcd2ff service nova] Acquiring lock "refresh_cache-31f9d37b-f4ff-4a8b-9477-8253f8305020" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.867392] env[61970]: DEBUG oslo_concurrency.lockutils [req-55eebb6b-935e-401d-b833-1b5d601f62de req-acbbd6cb-1a40-4c53-9595-c876c4bcd2ff service nova] Acquired lock "refresh_cache-31f9d37b-f4ff-4a8b-9477-8253f8305020" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.867635] env[61970]: DEBUG nova.network.neutron [req-55eebb6b-935e-401d-b833-1b5d601f62de req-acbbd6cb-1a40-4c53-9595-c876c4bcd2ff service nova] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Refreshing network info cache for port b39ddbde-1199-4f98-9fe2-cd05d19bc96e {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1014.985081] env[61970]: DEBUG oslo_concurrency.lockutils [None req-529409c3-5172-4b1a-8359-69e101ea5242 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "31f9d37b-f4ff-4a8b-9477-8253f8305020" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.700s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.051393] env[61970]: DEBUG oslo_vmware.api [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356052, 'name': PowerOffVM_Task, 'duration_secs': 0.490705} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.051763] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1015.052034] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Updating instance 'd38bde37-994b-4d58-b4b1-5f79b59fd2aa' progress to 17 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1015.104548] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': task-1356050, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.172485] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356051, 'name': Destroy_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.319934] env[61970]: DEBUG oslo_vmware.api [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356041, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.560698] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1015.560902] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1015.561082] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1015.561276] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1015.561486] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1015.561681] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1015.561984] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1015.562283] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1015.562626] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1015.562864] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1015.563067] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1015.571456] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-730af6b2-ce8c-489b-bfb1-6b38562e706d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.588699] env[61970]: DEBUG oslo_vmware.api [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 1015.588699] env[61970]: value = "task-1356053" [ 1015.588699] env[61970]: _type = "Task" [ 1015.588699] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.601232] env[61970]: DEBUG oslo_vmware.api [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356053, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.604367] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': task-1356050, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.072855} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.604831] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 19c0bfcb-1964-4719-98b1-d9334751ace2/19c0bfcb-1964-4719-98b1-d9334751ace2.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1015.605179] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1015.605179] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ad82b81e-0906-48bf-8927-847f4f5e8cc1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.611209] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Waiting for the task: (returnval){ [ 1015.611209] env[61970]: value = "task-1356054" [ 1015.611209] env[61970]: _type = "Task" [ 1015.611209] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.620136] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': task-1356054, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.662716] env[61970]: DEBUG nova.network.neutron [req-55eebb6b-935e-401d-b833-1b5d601f62de req-acbbd6cb-1a40-4c53-9595-c876c4bcd2ff service nova] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Updated VIF entry in instance network info cache for port b39ddbde-1199-4f98-9fe2-cd05d19bc96e. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1015.663093] env[61970]: DEBUG nova.network.neutron [req-55eebb6b-935e-401d-b833-1b5d601f62de req-acbbd6cb-1a40-4c53-9595-c876c4bcd2ff service nova] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Updating instance_info_cache with network_info: [{"id": "b39ddbde-1199-4f98-9fe2-cd05d19bc96e", "address": "fa:16:3e:be:c5:78", "network": {"id": "42def9ae-4eed-41e5-a538-81eaced2cdd4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2047751481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19dae2316adb4dc980bd972c8c40c034", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c1b8689-a9b4-4972-beb9-6a1c8de1dc88", "external-id": "nsx-vlan-transportzone-455", "segmentation_id": 455, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb39ddbde-11", "ovs_interfaceid": "b39ddbde-1199-4f98-9fe2-cd05d19bc96e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.671466] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356051, 'name': Destroy_Task, 'duration_secs': 1.016111} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.671729] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Destroyed the VM [ 1015.674030] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Deleting Snapshot of the VM instance {{(pid=61970) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1015.675229] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-18f97392-eb0d-43d5-8a14-5e70c6e5f0d5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.682113] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1015.682113] env[61970]: value = "task-1356055" [ 1015.682113] env[61970]: _type = "Task" [ 1015.682113] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.691594] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356055, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.819704] env[61970]: DEBUG oslo_vmware.api [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356041, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.101975] env[61970]: DEBUG oslo_vmware.api [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356053, 'name': ReconfigVM_Task, 'duration_secs': 0.41946} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.102389] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Updating instance 'd38bde37-994b-4d58-b4b1-5f79b59fd2aa' progress to 33 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1016.119760] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': task-1356054, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.267089} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.120086] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1016.120845] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f056c787-5623-48b8-9826-5f8cf0e21d69 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.143344] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 19c0bfcb-1964-4719-98b1-d9334751ace2/19c0bfcb-1964-4719-98b1-d9334751ace2.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1016.143629] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-644eb0c3-8ec2-44d9-ba97-ad3c1f0dd379 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.163398] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Waiting for the task: (returnval){ [ 1016.163398] env[61970]: value = "task-1356056" [ 1016.163398] env[61970]: _type = "Task" [ 1016.163398] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.166733] env[61970]: DEBUG oslo_concurrency.lockutils [req-55eebb6b-935e-401d-b833-1b5d601f62de req-acbbd6cb-1a40-4c53-9595-c876c4bcd2ff service nova] Releasing lock "refresh_cache-31f9d37b-f4ff-4a8b-9477-8253f8305020" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.171641] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': task-1356056, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.193603] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356055, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.320927] env[61970]: DEBUG oslo_vmware.api [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356041, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.609578] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1016.609886] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1016.610139] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1016.610446] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1016.610653] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1016.610841] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1016.611218] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1016.611494] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1016.611739] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1016.611994] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1016.612225] env[61970]: DEBUG nova.virt.hardware [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1016.618855] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Reconfiguring VM instance instance-00000060 to detach disk 2000 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1016.619231] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f68e648c-674f-42b4-8143-3376bc2bd46a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.638308] env[61970]: DEBUG oslo_vmware.api [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 1016.638308] env[61970]: value = "task-1356057" [ 1016.638308] env[61970]: _type = "Task" [ 1016.638308] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.647478] env[61970]: DEBUG oslo_vmware.api [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356057, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.672831] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': task-1356056, 'name': ReconfigVM_Task, 'duration_secs': 0.492797} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.673213] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 19c0bfcb-1964-4719-98b1-d9334751ace2/19c0bfcb-1964-4719-98b1-d9334751ace2.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1016.673914] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b1d9f6ee-665a-4b58-a0fe-acab3238e2b3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.682139] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Waiting for the task: (returnval){ [ 1016.682139] env[61970]: value = "task-1356058" [ 1016.682139] env[61970]: _type = "Task" [ 1016.682139] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.694751] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': task-1356058, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.698015] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356055, 'name': RemoveSnapshot_Task} progress is 65%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.821770] env[61970]: DEBUG oslo_vmware.api [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356041, 'name': ReconfigVM_Task, 'duration_secs': 6.291659} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.822044] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.822275] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Reconfigured VM to detach interface {{(pid=61970) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1017.148930] env[61970]: DEBUG oslo_vmware.api [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356057, 'name': ReconfigVM_Task, 'duration_secs': 0.215537} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.149243] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Reconfigured VM instance instance-00000060 to detach disk 2000 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1017.150032] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89353c21-4068-4a54-8cce-4bb8f973d2ed {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.171894] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] d38bde37-994b-4d58-b4b1-5f79b59fd2aa/d38bde37-994b-4d58-b4b1-5f79b59fd2aa.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1017.173542] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8d4200c9-ab9d-4c54-a856-21c5834d7201 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.200141] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': task-1356058, 'name': Rename_Task, 'duration_secs': 0.211617} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.203500] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1017.203881] env[61970]: DEBUG oslo_vmware.api [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356055, 'name': RemoveSnapshot_Task, 'duration_secs': 1.226913} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.205222] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b2dfc44f-ee46-4256-bf82-dedceb579bec {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.207741] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Deleted Snapshot of the VM instance {{(pid=61970) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1017.207741] env[61970]: INFO nova.compute.manager [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Took 19.94 seconds to snapshot the instance on the hypervisor. [ 1017.210873] env[61970]: DEBUG oslo_vmware.api [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 1017.210873] env[61970]: value = "task-1356059" [ 1017.210873] env[61970]: _type = "Task" [ 1017.210873] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.217768] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Waiting for the task: (returnval){ [ 1017.217768] env[61970]: value = "task-1356060" [ 1017.217768] env[61970]: _type = "Task" [ 1017.217768] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.222620] env[61970]: DEBUG oslo_vmware.api [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356059, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.231229] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': task-1356060, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.690759] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.691173] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquired lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.691173] env[61970]: DEBUG nova.network.neutron [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1017.729850] env[61970]: DEBUG oslo_vmware.api [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356059, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.735945] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': task-1356060, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.757923] env[61970]: DEBUG nova.compute.manager [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Found 3 images (rotation: 2) {{(pid=61970) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1017.758157] env[61970]: DEBUG nova.compute.manager [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Rotating out 1 backups {{(pid=61970) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 1017.758328] env[61970]: DEBUG nova.compute.manager [None req-a8470af9-74cc-4389-b157-4eacf9cf4ba4 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Deleting image 1a7f3a54-27a5-49f0-9a72-74cdc8f4f51b {{(pid=61970) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 1018.197113] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.197377] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.197628] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.197860] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.198063] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.204285] env[61970]: INFO nova.compute.manager [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Terminating instance [ 1018.207086] env[61970]: DEBUG nova.compute.manager [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1018.207286] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1018.208139] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a81ba96-641e-45f9-abc7-78ba74c76bf5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.216874] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1018.219766] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4ca14813-00a1-4b09-a2cd-45cbb2a9e7a0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.226859] env[61970]: DEBUG oslo_vmware.api [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356059, 'name': ReconfigVM_Task, 'duration_secs': 0.648163} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.232589] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Reconfigured VM instance instance-00000060 to attach disk [datastore2] d38bde37-994b-4d58-b4b1-5f79b59fd2aa/d38bde37-994b-4d58-b4b1-5f79b59fd2aa.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1018.232910] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Updating instance 'd38bde37-994b-4d58-b4b1-5f79b59fd2aa' progress to 50 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1018.236370] env[61970]: DEBUG oslo_vmware.api [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 1018.236370] env[61970]: value = "task-1356061" [ 1018.236370] env[61970]: _type = "Task" [ 1018.236370] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.243544] env[61970]: DEBUG oslo_vmware.api [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': task-1356060, 'name': PowerOnVM_Task, 'duration_secs': 0.722984} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.244286] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1018.244552] env[61970]: INFO nova.compute.manager [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Took 9.33 seconds to spawn the instance on the hypervisor. [ 1018.244795] env[61970]: DEBUG nova.compute.manager [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1018.245699] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5415601-92a5-4e95-91a7-659f74f4baaf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.251867] env[61970]: DEBUG oslo_vmware.api [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356061, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.491263] env[61970]: INFO nova.network.neutron [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Port 90fbe562-a7de-4ce8-a337-3c3f7be95120 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1018.491920] env[61970]: DEBUG nova.network.neutron [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Updating instance_info_cache with network_info: [{"id": "4434695d-ddda-4840-b1da-8159c98dcae0", "address": "fa:16:3e:85:27:97", "network": {"id": "8f622fc5-b74f-4582-a6e0-44c90f495750", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-67704340-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17ca05a1e6664430a02de563d98c1148", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4434695d-dd", "ovs_interfaceid": "4434695d-ddda-4840-b1da-8159c98dcae0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.739895] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f109ca46-4ca4-4fdb-af39-f3663e763dc8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.751289] env[61970]: DEBUG oslo_vmware.api [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356061, 'name': PowerOffVM_Task, 'duration_secs': 0.19515} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.763901] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1018.764107] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1018.769671] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-72c0fa55-0af3-40b1-bd6a-a882923b609f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.771628] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c2cc6c-15b5-40b8-9504-a70ce864c59b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.774092] env[61970]: INFO nova.compute.manager [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Took 14.99 seconds to build instance. [ 1018.793353] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Updating instance 'd38bde37-994b-4d58-b4b1-5f79b59fd2aa' progress to 67 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1018.848529] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1018.848759] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1018.848953] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Deleting the datastore file [datastore1] c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1018.849308] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8324eda9-9816-482e-994f-99e4da0d0f31 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.857773] env[61970]: DEBUG oslo_vmware.api [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 1018.857773] env[61970]: value = "task-1356063" [ 1018.857773] env[61970]: _type = "Task" [ 1018.857773] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.867818] env[61970]: DEBUG oslo_vmware.api [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356063, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.994459] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Releasing lock "refresh_cache-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.277116] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4a55ad3b-fdb0-4fe7-8c01-aa0240d2f7a6 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Lock "19c0bfcb-1964-4719-98b1-d9334751ace2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.504s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.338490] env[61970]: DEBUG nova.network.neutron [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Port 62d22328-b863-41dc-816f-1ce3f7a53eed binding to destination host cpu-1 is already ACTIVE {{(pid=61970) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1019.369596] env[61970]: DEBUG oslo_vmware.api [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356063, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143758} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.369858] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1019.370112] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1019.370321] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1019.370500] env[61970]: INFO nova.compute.manager [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1019.370740] env[61970]: DEBUG oslo.service.loopingcall [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1019.370933] env[61970]: DEBUG nova.compute.manager [-] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1019.371037] env[61970]: DEBUG nova.network.neutron [-] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1019.499557] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e5716b58-f9d9-478f-82b8-51296f382329 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "interface-c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8-90fbe562-a7de-4ce8-a337-3c3f7be95120" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.800s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.693286] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.693535] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.847373] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dcdd3e67-5f90-49dc-99a6-de68e5368d47 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "ebb92a5b-f635-4504-a57f-395de514015a" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.847690] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dcdd3e67-5f90-49dc-99a6-de68e5368d47 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "ebb92a5b-f635-4504-a57f-395de514015a" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.847788] env[61970]: DEBUG nova.compute.manager [None req-dcdd3e67-5f90-49dc-99a6-de68e5368d47 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1019.848866] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b53853c-bd0c-4c54-b164-cd569485b443 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.856369] env[61970]: DEBUG nova.compute.manager [None req-dcdd3e67-5f90-49dc-99a6-de68e5368d47 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61970) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1019.856921] env[61970]: DEBUG nova.objects.instance [None req-dcdd3e67-5f90-49dc-99a6-de68e5368d47 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lazy-loading 'flavor' on Instance uuid ebb92a5b-f635-4504-a57f-395de514015a {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1020.145810] env[61970]: DEBUG nova.compute.manager [req-e7c853d4-9a6f-4967-8824-a2904cb456b8 req-41caaa41-8d93-4f80-8533-54710457655e service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Received event network-vif-deleted-4434695d-ddda-4840-b1da-8159c98dcae0 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.146028] env[61970]: INFO nova.compute.manager [req-e7c853d4-9a6f-4967-8824-a2904cb456b8 req-41caaa41-8d93-4f80-8533-54710457655e service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Neutron deleted interface 4434695d-ddda-4840-b1da-8159c98dcae0; detaching it from the instance and deleting it from the info cache [ 1020.146613] env[61970]: DEBUG nova.network.neutron [req-e7c853d4-9a6f-4967-8824-a2904cb456b8 req-41caaa41-8d93-4f80-8533-54710457655e service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.196726] env[61970]: DEBUG nova.compute.utils [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1020.366755] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcdd3e67-5f90-49dc-99a6-de68e5368d47 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1020.367590] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "d38bde37-994b-4d58-b4b1-5f79b59fd2aa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.367817] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "d38bde37-994b-4d58-b4b1-5f79b59fd2aa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.367995] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "d38bde37-994b-4d58-b4b1-5f79b59fd2aa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.372021] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dabf18d9-b9ec-4051-943f-c646850ffeee {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.374423] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Acquiring lock "19c0bfcb-1964-4719-98b1-d9334751ace2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.374939] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Lock "19c0bfcb-1964-4719-98b1-d9334751ace2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.375223] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Acquiring lock "19c0bfcb-1964-4719-98b1-d9334751ace2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.375417] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Lock "19c0bfcb-1964-4719-98b1-d9334751ace2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.375610] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Lock "19c0bfcb-1964-4719-98b1-d9334751ace2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.378797] env[61970]: DEBUG oslo_vmware.api [None req-dcdd3e67-5f90-49dc-99a6-de68e5368d47 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1020.378797] env[61970]: value = "task-1356064" [ 1020.378797] env[61970]: _type = "Task" [ 1020.378797] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.379748] env[61970]: INFO nova.compute.manager [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Terminating instance [ 1020.381732] env[61970]: DEBUG nova.compute.manager [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1020.382476] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1020.386248] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cb3453d-c0d4-46b1-bc33-dac18d36b892 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.398616] env[61970]: DEBUG oslo_vmware.api [None req-dcdd3e67-5f90-49dc-99a6-de68e5368d47 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356064, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.401412] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1020.401573] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-12e3d4a6-4d29-4e24-93b8-ee4718eb55e1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.409726] env[61970]: DEBUG oslo_vmware.api [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Waiting for the task: (returnval){ [ 1020.409726] env[61970]: value = "task-1356065" [ 1020.409726] env[61970]: _type = "Task" [ 1020.409726] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.420458] env[61970]: DEBUG oslo_vmware.api [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': task-1356065, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.626665] env[61970]: DEBUG nova.network.neutron [-] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.649273] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-512a8a31-68d4-4e52-b97b-73bba4e941b0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.659182] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc07140-4e32-4f61-b0c3-5b2d5deb6139 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.690204] env[61970]: DEBUG nova.compute.manager [req-e7c853d4-9a6f-4967-8824-a2904cb456b8 req-41caaa41-8d93-4f80-8533-54710457655e service nova] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Detach interface failed, port_id=4434695d-ddda-4840-b1da-8159c98dcae0, reason: Instance c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1020.699099] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.892123] env[61970]: DEBUG oslo_vmware.api [None req-dcdd3e67-5f90-49dc-99a6-de68e5368d47 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356064, 'name': PowerOffVM_Task, 'duration_secs': 0.195376} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.892429] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcdd3e67-5f90-49dc-99a6-de68e5368d47 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1020.892607] env[61970]: DEBUG nova.compute.manager [None req-dcdd3e67-5f90-49dc-99a6-de68e5368d47 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1020.893366] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017caafe-31de-4155-bc72-d3aa5b8f8d1b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.919794] env[61970]: DEBUG oslo_vmware.api [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': task-1356065, 'name': PowerOffVM_Task, 'duration_secs': 0.214174} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.920089] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1020.920301] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1020.920557] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-98c0c09d-da9f-4d9a-8a72-663d6c2c565e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.988083] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1020.988397] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1020.988601] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Deleting the datastore file [datastore2] 19c0bfcb-1964-4719-98b1-d9334751ace2 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1020.988879] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-acef028d-fe21-4093-9fd8-884949cdefae {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.996667] env[61970]: DEBUG oslo_vmware.api [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Waiting for the task: (returnval){ [ 1020.996667] env[61970]: value = "task-1356067" [ 1020.996667] env[61970]: _type = "Task" [ 1020.996667] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.006057] env[61970]: DEBUG oslo_vmware.api [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': task-1356067, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.129688] env[61970]: INFO nova.compute.manager [-] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Took 1.76 seconds to deallocate network for instance. [ 1021.404927] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dcdd3e67-5f90-49dc-99a6-de68e5368d47 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "ebb92a5b-f635-4504-a57f-395de514015a" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.557s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.406262] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "refresh_cache-d38bde37-994b-4d58-b4b1-5f79b59fd2aa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.406433] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "refresh_cache-d38bde37-994b-4d58-b4b1-5f79b59fd2aa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.406610] env[61970]: DEBUG nova.network.neutron [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1021.508643] env[61970]: DEBUG oslo_vmware.api [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Task: {'id': task-1356067, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183176} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.508994] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1021.509296] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1021.509557] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1021.509757] env[61970]: INFO nova.compute.manager [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1021.510018] env[61970]: DEBUG oslo.service.loopingcall [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1021.510240] env[61970]: DEBUG nova.compute.manager [-] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1021.510355] env[61970]: DEBUG nova.network.neutron [-] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1021.636919] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.637337] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.637633] env[61970]: DEBUG nova.objects.instance [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lazy-loading 'resources' on Instance uuid c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1021.763762] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.764059] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.764312] env[61970]: INFO nova.compute.manager [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Attaching volume 8724c3ef-36fc-44b6-a9e3-dfee76f52fa2 to /dev/sdb [ 1021.802809] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7bbc96a-1af9-4ac2-97cc-b054f4a595f2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.810879] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bfcab33-3670-482c-bef2-4109a22df87b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.828864] env[61970]: DEBUG nova.virt.block_device [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Updating existing volume attachment record: da6cd478-53ba-46ae-a628-ec18dfdd16a9 {{(pid=61970) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1022.174390] env[61970]: DEBUG nova.compute.manager [req-571548c4-80ff-4fd7-ac6b-ee7081574bfa req-f727fd3f-823c-4728-a93d-09ca54121e4d service nova] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Received event network-vif-deleted-babd8ae5-2695-4eb8-97d1-e9db1f6c518e {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1022.174878] env[61970]: INFO nova.compute.manager [req-571548c4-80ff-4fd7-ac6b-ee7081574bfa req-f727fd3f-823c-4728-a93d-09ca54121e4d service nova] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Neutron deleted interface babd8ae5-2695-4eb8-97d1-e9db1f6c518e; detaching it from the instance and deleting it from the info cache [ 1022.174878] env[61970]: DEBUG nova.network.neutron [req-571548c4-80ff-4fd7-ac6b-ee7081574bfa req-f727fd3f-823c-4728-a93d-09ca54121e4d service nova] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.180231] env[61970]: DEBUG nova.network.neutron [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Updating instance_info_cache with network_info: [{"id": "62d22328-b863-41dc-816f-1ce3f7a53eed", "address": "fa:16:3e:53:70:05", "network": {"id": "62edd125-b77c-4418-a465-1dcb3a0733c1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-371307366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6a4997482a4834a30c79c4cbab2cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62d22328-b8", "ovs_interfaceid": "62d22328-b863-41dc-816f-1ce3f7a53eed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.297654] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0f69d97-3779-44f4-8719-3964251c26c7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.301937] env[61970]: DEBUG nova.network.neutron [-] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.309250] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bfb08c3-a159-454a-ada4-52d2e4de5773 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.342554] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b816a34-c091-4bdb-8483-e779416f2f49 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.352030] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-208b5e05-e983-4cd8-ac60-f0ab08c8cfe1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.367393] env[61970]: DEBUG nova.compute.provider_tree [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.679611] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1472f9d7-db96-42f4-b6ba-fe1fd11bdc4d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.682284] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "refresh_cache-d38bde37-994b-4d58-b4b1-5f79b59fd2aa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.689501] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22c10b3c-a001-41bb-ae02-7792c2ae1ffb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.718452] env[61970]: DEBUG nova.compute.manager [req-571548c4-80ff-4fd7-ac6b-ee7081574bfa req-f727fd3f-823c-4728-a93d-09ca54121e4d service nova] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Detach interface failed, port_id=babd8ae5-2695-4eb8-97d1-e9db1f6c518e, reason: Instance 19c0bfcb-1964-4719-98b1-d9334751ace2 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1022.804937] env[61970]: INFO nova.compute.manager [-] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Took 1.29 seconds to deallocate network for instance. [ 1022.870166] env[61970]: DEBUG nova.scheduler.client.report [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1023.038665] env[61970]: DEBUG nova.compute.manager [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Stashing vm_state: stopped {{(pid=61970) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1023.207385] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3705d62-bcee-4c0d-9e89-49ba1f9f9b40 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.228167] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2743ecc5-b428-4cc6-a262-58c9f7ba6e69 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.236435] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Updating instance 'd38bde37-994b-4d58-b4b1-5f79b59fd2aa' progress to 83 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1023.311672] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.375563] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.738s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.377867] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.066s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.378270] env[61970]: DEBUG nova.objects.instance [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Lazy-loading 'resources' on Instance uuid 19c0bfcb-1964-4719-98b1-d9334751ace2 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1023.396441] env[61970]: INFO nova.scheduler.client.report [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Deleted allocations for instance c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8 [ 1023.555245] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.742958] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1023.743289] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c80a8304-911d-49f6-bd9b-855a8faee2ac {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.752718] env[61970]: DEBUG oslo_vmware.api [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 1023.752718] env[61970]: value = "task-1356071" [ 1023.752718] env[61970]: _type = "Task" [ 1023.752718] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.760335] env[61970]: DEBUG oslo_vmware.api [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356071, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.905105] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e702be1-d8c6-4359-92c5-7a3145230ca5 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.708s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.016072] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5205d9ca-4ac3-43fb-845b-5194c7f9e21f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.025024] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24ec84c1-c05e-456f-aa6a-89250956642d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.056832] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a95bab-232c-45c9-94cd-9d53b771b1cc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.064867] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fdc92ee-6ee5-4c00-92d9-bdd091a94db0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.079292] env[61970]: DEBUG nova.compute.provider_tree [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.262870] env[61970]: DEBUG oslo_vmware.api [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356071, 'name': PowerOnVM_Task, 'duration_secs': 0.387447} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.263258] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1024.263336] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0c036d7f-4c70-4cf2-9081-db55c7a90003 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Updating instance 'd38bde37-994b-4d58-b4b1-5f79b59fd2aa' progress to 100 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1024.583061] env[61970]: DEBUG nova.scheduler.client.report [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1024.923776] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.924058] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.924305] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.924955] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.925183] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.927361] env[61970]: INFO nova.compute.manager [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Terminating instance [ 1024.929199] env[61970]: DEBUG nova.compute.manager [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1024.929425] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1024.930278] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa9504b-147f-4881-88ed-cd027add5c37 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.938855] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1024.939097] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c23ca4a9-79f2-4b8c-8d45-96e2e9a09eaf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.946649] env[61970]: DEBUG oslo_vmware.api [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 1024.946649] env[61970]: value = "task-1356073" [ 1024.946649] env[61970]: _type = "Task" [ 1024.946649] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.955131] env[61970]: DEBUG oslo_vmware.api [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356073, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.089120] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.711s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.092150] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.537s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.112114] env[61970]: INFO nova.scheduler.client.report [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Deleted allocations for instance 19c0bfcb-1964-4719-98b1-d9334751ace2 [ 1025.459379] env[61970]: DEBUG oslo_vmware.api [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356073, 'name': PowerOffVM_Task, 'duration_secs': 0.177151} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.459379] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1025.459379] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1025.459379] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-83e5ec06-5056-4a72-b34b-6879ed1ab088 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.524634] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1025.525077] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1025.525200] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Deleting the datastore file [datastore1] c4ec1797-d939-4c24-9314-ce9c74eb5aa1 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1025.525455] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4d83ddaa-3b7b-45ec-8266-469115527e49 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.531793] env[61970]: DEBUG oslo_vmware.api [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for the task: (returnval){ [ 1025.531793] env[61970]: value = "task-1356075" [ 1025.531793] env[61970]: _type = "Task" [ 1025.531793] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.540373] env[61970]: DEBUG oslo_vmware.api [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356075, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.598021] env[61970]: INFO nova.compute.claims [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1025.619705] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3d645ff4-ca8c-4215-b027-87a6d7bc4691 tempest-ServerTagsTestJSON-916334356 tempest-ServerTagsTestJSON-916334356-project-member] Lock "19c0bfcb-1964-4719-98b1-d9334751ace2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.245s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.042876] env[61970]: DEBUG oslo_vmware.api [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Task: {'id': task-1356075, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144684} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.043118] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1026.043314] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1026.043548] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1026.043826] env[61970]: INFO nova.compute.manager [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1026.043997] env[61970]: DEBUG oslo.service.loopingcall [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.044216] env[61970]: DEBUG nova.compute.manager [-] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1026.044313] env[61970]: DEBUG nova.network.neutron [-] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1026.105861] env[61970]: INFO nova.compute.resource_tracker [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updating resource usage from migration e20299e1-a255-41de-9c03-b97e0ba982fd [ 1026.293676] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-100e9693-b3de-496a-aa15-82cdc7d824d7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.305891] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9a38215-3e62-4f27-9c4b-e07947277c4f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.316814] env[61970]: DEBUG nova.compute.manager [req-990f296e-ca84-4eda-a896-2d0a8ca54372 req-aca99883-fefc-4773-831d-deae89cfc5d7 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Received event network-vif-deleted-5b506f12-7feb-41e2-8e18-abc9262ac87b {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.317098] env[61970]: INFO nova.compute.manager [req-990f296e-ca84-4eda-a896-2d0a8ca54372 req-aca99883-fefc-4773-831d-deae89cfc5d7 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Neutron deleted interface 5b506f12-7feb-41e2-8e18-abc9262ac87b; detaching it from the instance and deleting it from the info cache [ 1026.317342] env[61970]: DEBUG nova.network.neutron [req-990f296e-ca84-4eda-a896-2d0a8ca54372 req-aca99883-fefc-4773-831d-deae89cfc5d7 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.351340] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d56d47-bc28-4ad1-9f46-b60f9ba41c79 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.352076] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3e5f2937-b895-49f1-842b-da66555bcb63 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.362319] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-604228f1-99cb-4ff9-a014-f5ed650fe05d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.370012] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc9c6410-5db6-4e10-8d88-7d73c1663f65 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.382485] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Volume attach. Driver type: vmdk {{(pid=61970) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1026.382788] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288868', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'name': 'volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bc948a41-1a22-46fa-a76a-50b4c4afa4b7', 'attached_at': '', 'detached_at': '', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'serial': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1026.384611] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232cc68b-6157-4136-aded-9f580020f3e6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.396561] env[61970]: DEBUG nova.compute.provider_tree [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.421132] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f408d777-4c8a-421c-9284-18fd9a95866b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.424776] env[61970]: DEBUG nova.compute.manager [req-990f296e-ca84-4eda-a896-2d0a8ca54372 req-aca99883-fefc-4773-831d-deae89cfc5d7 service nova] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Detach interface failed, port_id=5b506f12-7feb-41e2-8e18-abc9262ac87b, reason: Instance c4ec1797-d939-4c24-9314-ce9c74eb5aa1 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1026.451028] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2/volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1026.451028] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a4e3364-70f0-4025-80f9-fa1a649af247 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.472095] env[61970]: DEBUG oslo_vmware.api [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1026.472095] env[61970]: value = "task-1356076" [ 1026.472095] env[61970]: _type = "Task" [ 1026.472095] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.482269] env[61970]: DEBUG oslo_vmware.api [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356076, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.786142] env[61970]: DEBUG nova.network.neutron [-] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.899775] env[61970]: DEBUG nova.scheduler.client.report [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1026.983387] env[61970]: DEBUG oslo_vmware.api [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356076, 'name': ReconfigVM_Task, 'duration_secs': 0.382611} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.983702] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Reconfigured VM instance instance-0000005d to attach disk [datastore1] volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2/volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1026.988527] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1ebc8b4-ec6d-441e-87b8-a21b791e823c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.005070] env[61970]: DEBUG oslo_vmware.api [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1027.005070] env[61970]: value = "task-1356077" [ 1027.005070] env[61970]: _type = "Task" [ 1027.005070] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.014066] env[61970]: DEBUG oslo_vmware.api [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356077, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.073403] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "d38bde37-994b-4d58-b4b1-5f79b59fd2aa" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.073765] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "d38bde37-994b-4d58-b4b1-5f79b59fd2aa" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.073987] env[61970]: DEBUG nova.compute.manager [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Going to confirm migration 3 {{(pid=61970) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1027.289261] env[61970]: INFO nova.compute.manager [-] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Took 1.24 seconds to deallocate network for instance. [ 1027.405530] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.313s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.405836] env[61970]: INFO nova.compute.manager [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Migrating [ 1027.515895] env[61970]: DEBUG oslo_vmware.api [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356077, 'name': ReconfigVM_Task, 'duration_secs': 0.14698} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.516243] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288868', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'name': 'volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bc948a41-1a22-46fa-a76a-50b4c4afa4b7', 'attached_at': '', 'detached_at': '', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'serial': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1027.625887] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "refresh_cache-d38bde37-994b-4d58-b4b1-5f79b59fd2aa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.626112] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquired lock "refresh_cache-d38bde37-994b-4d58-b4b1-5f79b59fd2aa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.626313] env[61970]: DEBUG nova.network.neutron [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1027.626506] env[61970]: DEBUG nova.objects.instance [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lazy-loading 'info_cache' on Instance uuid d38bde37-994b-4d58-b4b1-5f79b59fd2aa {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.796564] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.796844] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.797082] env[61970]: DEBUG nova.objects.instance [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lazy-loading 'resources' on Instance uuid c4ec1797-d939-4c24-9314-ce9c74eb5aa1 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.927124] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.927310] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.927505] env[61970]: DEBUG nova.network.neutron [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1028.433636] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-287ed446-fcbf-44da-a02b-7ff4ba641894 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.443221] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeec0a6c-7956-45dc-a14e-fd09c0741dc1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.476073] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe1af76-ef98-47ef-82ec-26cbf94d2201 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.483953] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc4ee016-0c48-49b0-b0bd-3d6f40c1f921 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.497689] env[61970]: DEBUG nova.compute.provider_tree [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.561887] env[61970]: DEBUG nova.objects.instance [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lazy-loading 'flavor' on Instance uuid bc948a41-1a22-46fa-a76a-50b4c4afa4b7 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1028.786550] env[61970]: DEBUG nova.network.neutron [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updating instance_info_cache with network_info: [{"id": "9a1c8b46-0249-4e0b-94de-b0d0062ca69c", "address": "fa:16:3e:93:5e:22", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a1c8b46-02", "ovs_interfaceid": "9a1c8b46-0249-4e0b-94de-b0d0062ca69c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.000293] env[61970]: DEBUG nova.scheduler.client.report [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.066057] env[61970]: DEBUG oslo_concurrency.lockutils [None req-fb3e93b2-8e13-48b3-9e28-766a28560b42 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.302s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.130527] env[61970]: DEBUG nova.network.neutron [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Updating instance_info_cache with network_info: [{"id": "62d22328-b863-41dc-816f-1ce3f7a53eed", "address": "fa:16:3e:53:70:05", "network": {"id": "62edd125-b77c-4418-a465-1dcb3a0733c1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-371307366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6a4997482a4834a30c79c4cbab2cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62d22328-b8", "ovs_interfaceid": "62d22328-b863-41dc-816f-1ce3f7a53eed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.292586] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.508703] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.712s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.530826] env[61970]: INFO nova.scheduler.client.report [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Deleted allocations for instance c4ec1797-d939-4c24-9314-ce9c74eb5aa1 [ 1029.633067] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Releasing lock "refresh_cache-d38bde37-994b-4d58-b4b1-5f79b59fd2aa" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.633434] env[61970]: DEBUG nova.objects.instance [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lazy-loading 'migration_context' on Instance uuid d38bde37-994b-4d58-b4b1-5f79b59fd2aa {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.753720] env[61970]: INFO nova.compute.manager [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Rebuilding instance [ 1029.792918] env[61970]: DEBUG nova.compute.manager [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1029.793933] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e1f332f-b556-4b67-90e7-4fb94193215e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.039261] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df38acaa-e950-4c2a-bd81-93d37babc180 tempest-AttachInterfacesTestJSON-55273928 tempest-AttachInterfacesTestJSON-55273928-project-member] Lock "c4ec1797-d939-4c24-9314-ce9c74eb5aa1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.115s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.135912] env[61970]: DEBUG nova.objects.base [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1030.136829] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-997055b8-4796-48a4-aaca-9849ca2b19f2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.158835] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07325099-3ef1-4c4e-8a24-a942f787ca2b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.168753] env[61970]: DEBUG oslo_vmware.api [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 1030.168753] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ebd346-209e-cbb3-204b-5d1b3da85190" [ 1030.168753] env[61970]: _type = "Task" [ 1030.168753] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.181142] env[61970]: DEBUG oslo_vmware.api [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ebd346-209e-cbb3-204b-5d1b3da85190, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.308902] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1030.308902] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7da1c1e4-af44-4c55-8401-53b7f4c51de6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.316165] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1030.316165] env[61970]: value = "task-1356078" [ 1030.316165] env[61970]: _type = "Task" [ 1030.316165] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.325051] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356078, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.679653] env[61970]: DEBUG oslo_vmware.api [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ebd346-209e-cbb3-204b-5d1b3da85190, 'name': SearchDatastore_Task, 'duration_secs': 0.009974} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.680109] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.680262] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.808983] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aeabc79-5b02-407c-8473-736254cd6159 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.831424] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updating instance 'ebb92a5b-f635-4504-a57f-395de514015a' progress to 0 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1030.843338] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356078, 'name': PowerOffVM_Task, 'duration_secs': 0.241561} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.843691] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1030.904207] env[61970]: INFO nova.compute.manager [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Detaching volume 8724c3ef-36fc-44b6-a9e3-dfee76f52fa2 [ 1030.936550] env[61970]: INFO nova.virt.block_device [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Attempting to driver detach volume 8724c3ef-36fc-44b6-a9e3-dfee76f52fa2 from mountpoint /dev/sdb [ 1030.936984] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Volume detach. Driver type: vmdk {{(pid=61970) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1030.936984] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288868', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'name': 'volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bc948a41-1a22-46fa-a76a-50b4c4afa4b7', 'attached_at': '', 'detached_at': '', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'serial': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1030.938027] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bdd1218-d67c-4075-8664-6bfccd3f8bcd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.960078] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e04704-06b2-4f24-aec5-2fd6b90f3866 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.968249] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00370f06-8c73-4588-a278-fde7083f756d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.989149] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8e05ac-ddc7-407a-8327-9996ef8bd0c2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.007441] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] The volume has not been displaced from its original location: [datastore1] volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2/volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2.vmdk. No consolidation needed. {{(pid=61970) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1031.012764] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Reconfiguring VM instance instance-0000005d to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1031.013105] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d093bff8-f232-43ac-b1ed-06d4642ebe6a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.032770] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1031.032770] env[61970]: value = "task-1356079" [ 1031.032770] env[61970]: _type = "Task" [ 1031.032770] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.041124] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356079, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.340870] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1031.341513] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-687d778e-cf31-4f14-b76d-dd1680fc7c30 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.350371] env[61970]: DEBUG oslo_vmware.api [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1031.350371] env[61970]: value = "task-1356080" [ 1031.350371] env[61970]: _type = "Task" [ 1031.350371] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.362810] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] VM already powered off {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1031.362810] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updating instance 'ebb92a5b-f635-4504-a57f-395de514015a' progress to 17 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1031.418324] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5311c035-b90d-4268-9056-b561eca8877f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.427416] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b7aa89e-f27a-4604-94e3-ce18b2fc5422 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.458646] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4718c3f9-4e99-4761-81bc-57ba9f7ec4c1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.467216] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61ef83bc-357d-4992-a8fd-c4b006e46079 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.481293] env[61970]: DEBUG nova.compute.provider_tree [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1031.545191] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356079, 'name': ReconfigVM_Task, 'duration_secs': 0.202201} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.545482] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Reconfigured VM instance instance-0000005d to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1031.549996] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd6fc2b8-5b0a-4f92-bada-a92f51fde1f7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.566467] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1031.566467] env[61970]: value = "task-1356081" [ 1031.566467] env[61970]: _type = "Task" [ 1031.566467] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.574831] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356081, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.874023] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1031.874023] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1031.874023] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1031.874023] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1031.874023] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1031.874023] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1031.874023] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1031.874023] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1031.874023] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1031.874023] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1031.874023] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.878647] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9197766b-36c2-4a0b-8cfb-ae7344af70d6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.896579] env[61970]: DEBUG oslo_vmware.api [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1031.896579] env[61970]: value = "task-1356082" [ 1031.896579] env[61970]: _type = "Task" [ 1031.896579] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.906282] env[61970]: DEBUG oslo_vmware.api [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356082, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.984507] env[61970]: DEBUG nova.scheduler.client.report [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.079177] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356081, 'name': ReconfigVM_Task, 'duration_secs': 0.511268} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.079490] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288868', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'name': 'volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bc948a41-1a22-46fa-a76a-50b4c4afa4b7', 'attached_at': '', 'detached_at': '', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'serial': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1032.409782] env[61970]: DEBUG oslo_vmware.api [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356082, 'name': ReconfigVM_Task, 'duration_secs': 0.211944} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.410132] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updating instance 'ebb92a5b-f635-4504-a57f-395de514015a' progress to 33 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1032.917100] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1032.917451] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1032.917528] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1032.917711] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1032.917879] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1032.918141] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1032.918395] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1032.918550] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1032.918723] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1032.918889] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1032.919077] env[61970]: DEBUG nova.virt.hardware [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1032.924421] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Reconfiguring VM instance instance-00000051 to detach disk 2000 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1032.924716] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae70708a-d49f-493d-b8db-8ed2e013104f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.944223] env[61970]: DEBUG oslo_vmware.api [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1032.944223] env[61970]: value = "task-1356083" [ 1032.944223] env[61970]: _type = "Task" [ 1032.944223] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.952762] env[61970]: DEBUG oslo_vmware.api [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356083, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.998361] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.318s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.136507] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1033.136843] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c932cb80-10dc-42d0-84dc-6bf437738aaf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.144669] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1033.144669] env[61970]: value = "task-1356084" [ 1033.144669] env[61970]: _type = "Task" [ 1033.144669] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.153430] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356084, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.454315] env[61970]: DEBUG oslo_vmware.api [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356083, 'name': ReconfigVM_Task, 'duration_secs': 0.16313} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.454595] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Reconfigured VM instance instance-00000051 to detach disk 2000 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1033.455381] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd0a38d4-d72c-4e95-a2b0-e36f8ce17419 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.477231] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] ebb92a5b-f635-4504-a57f-395de514015a/ebb92a5b-f635-4504-a57f-395de514015a.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1033.477506] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10203a1e-89a3-4b25-98d6-3be9755b6301 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.496486] env[61970]: DEBUG oslo_vmware.api [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1033.496486] env[61970]: value = "task-1356085" [ 1033.496486] env[61970]: _type = "Task" [ 1033.496486] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.508761] env[61970]: DEBUG oslo_vmware.api [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356085, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.575139] env[61970]: INFO nova.scheduler.client.report [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Deleted allocation for migration 63900b78-7846-41b5-89aa-ce6edab35b4e [ 1033.658336] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] VM already powered off {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1033.658574] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Volume detach. Driver type: vmdk {{(pid=61970) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1033.658892] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288868', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'name': 'volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bc948a41-1a22-46fa-a76a-50b4c4afa4b7', 'attached_at': '', 'detached_at': '', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'serial': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1033.659709] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-631db187-14ff-471b-873c-704e42b778b2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.679349] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cfb06a7-1ca5-4c34-947c-78b8d9d80b89 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.686630] env[61970]: WARNING nova.virt.vmwareapi.driver [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1033.686941] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1033.687695] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-382f6a9a-b86c-4c7c-92d0-3cf117762509 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.695094] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1033.695342] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eddf7cf1-74b7-49e6-8b03-1b379a6ac7f1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.756712] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1033.756951] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1033.757277] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleting the datastore file [datastore2] bc948a41-1a22-46fa-a76a-50b4c4afa4b7 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1033.757429] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-02083e1f-fee3-4e6e-bc4e-f55bef6989ff {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.765564] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1033.765564] env[61970]: value = "task-1356087" [ 1033.765564] env[61970]: _type = "Task" [ 1033.765564] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.773823] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356087, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.008418] env[61970]: DEBUG oslo_vmware.api [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356085, 'name': ReconfigVM_Task, 'duration_secs': 0.256516} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.008713] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Reconfigured VM instance instance-00000051 to attach disk [datastore2] ebb92a5b-f635-4504-a57f-395de514015a/ebb92a5b-f635-4504-a57f-395de514015a.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1034.009016] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updating instance 'ebb92a5b-f635-4504-a57f-395de514015a' progress to 50 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1034.081523] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "d38bde37-994b-4d58-b4b1-5f79b59fd2aa" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.008s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.275946] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356087, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146741} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.276244] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1034.276440] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1034.276619] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1034.516471] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec168782-5565-4db3-a849-2c1b02c0c42a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.537491] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d125765d-df0b-445b-8d9b-389748ef043e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.560427] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updating instance 'ebb92a5b-f635-4504-a57f-395de514015a' progress to 67 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1034.782178] env[61970]: INFO nova.virt.block_device [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Booting with volume 8724c3ef-36fc-44b6-a9e3-dfee76f52fa2 at /dev/sdb [ 1034.826765] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f526947d-10f1-4f47-b175-09037e83d84d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.838745] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d78ba4bd-5c7d-4ada-b5ad-668c1c661d31 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.872183] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-40d9dcbf-b781-4779-aa63-103de3a165b0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.883058] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8024cc70-ff06-4ffd-9b6b-bd4f6fad927f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.918216] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e22a00c-af3e-4a80-ad18-51b50532bc1f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.926062] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9609e589-c970-44e9-b363-5c165cd99a85 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.941570] env[61970]: DEBUG nova.virt.block_device [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Updating existing volume attachment record: 082ec49d-1478-459d-a658-076688752682 {{(pid=61970) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1035.110539] env[61970]: DEBUG nova.network.neutron [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Port 9a1c8b46-0249-4e0b-94de-b0d0062ca69c binding to destination host cpu-1 is already ACTIVE {{(pid=61970) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1035.613525] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "d38bde37-994b-4d58-b4b1-5f79b59fd2aa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.613808] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "d38bde37-994b-4d58-b4b1-5f79b59fd2aa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.614050] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "d38bde37-994b-4d58-b4b1-5f79b59fd2aa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.614247] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "d38bde37-994b-4d58-b4b1-5f79b59fd2aa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.614427] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "d38bde37-994b-4d58-b4b1-5f79b59fd2aa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.621351] env[61970]: INFO nova.compute.manager [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Terminating instance [ 1035.623673] env[61970]: DEBUG nova.compute.manager [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1035.624321] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1035.624860] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7859bfb-369e-4521-973d-b127d0ef1417 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.633560] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1035.633803] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5db7c077-71c6-42b1-9c0c-a59b80cf1730 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.645021] env[61970]: DEBUG oslo_vmware.api [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 1035.645021] env[61970]: value = "task-1356088" [ 1035.645021] env[61970]: _type = "Task" [ 1035.645021] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.650979] env[61970]: DEBUG oslo_vmware.api [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356088, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.142176] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "ebb92a5b-f635-4504-a57f-395de514015a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.142506] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "ebb92a5b-f635-4504-a57f-395de514015a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.142606] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "ebb92a5b-f635-4504-a57f-395de514015a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.157288] env[61970]: DEBUG oslo_vmware.api [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356088, 'name': PowerOffVM_Task, 'duration_secs': 0.254329} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.157288] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1036.157288] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1036.158031] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c04f8e65-c89a-4ddb-ad6c-6cbbe9ebe496 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.239363] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1036.239635] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1036.239917] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Deleting the datastore file [datastore2] d38bde37-994b-4d58-b4b1-5f79b59fd2aa {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1036.240110] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-457d4438-e60f-4e94-8300-c8220019c576 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.249919] env[61970]: DEBUG oslo_vmware.api [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for the task: (returnval){ [ 1036.249919] env[61970]: value = "task-1356090" [ 1036.249919] env[61970]: _type = "Task" [ 1036.249919] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.258381] env[61970]: DEBUG oslo_vmware.api [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356090, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.765924] env[61970]: DEBUG oslo_vmware.api [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Task: {'id': task-1356090, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146836} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.766436] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1036.766608] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1036.766897] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1036.767192] env[61970]: INFO nova.compute.manager [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1036.767704] env[61970]: DEBUG oslo.service.loopingcall [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.767842] env[61970]: DEBUG nova.compute.manager [-] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1036.768041] env[61970]: DEBUG nova.network.neutron [-] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1037.005194] env[61970]: DEBUG nova.compute.manager [req-9eafe46b-daa4-404a-ae9a-35f73b93ade4 req-5202f602-aa18-4218-bb35-fcc7a3aab29a service nova] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Received event network-vif-deleted-62d22328-b863-41dc-816f-1ce3f7a53eed {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1037.007706] env[61970]: INFO nova.compute.manager [req-9eafe46b-daa4-404a-ae9a-35f73b93ade4 req-5202f602-aa18-4218-bb35-fcc7a3aab29a service nova] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Neutron deleted interface 62d22328-b863-41dc-816f-1ce3f7a53eed; detaching it from the instance and deleting it from the info cache [ 1037.007706] env[61970]: DEBUG nova.network.neutron [req-9eafe46b-daa4-404a-ae9a-35f73b93ade4 req-5202f602-aa18-4218-bb35-fcc7a3aab29a service nova] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.068298] env[61970]: DEBUG nova.virt.hardware [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1037.068298] env[61970]: DEBUG nova.virt.hardware [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1037.068594] env[61970]: DEBUG nova.virt.hardware [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1037.068594] env[61970]: DEBUG nova.virt.hardware [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1037.068734] env[61970]: DEBUG nova.virt.hardware [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1037.068886] env[61970]: DEBUG nova.virt.hardware [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1037.069106] env[61970]: DEBUG nova.virt.hardware [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1037.069353] env[61970]: DEBUG nova.virt.hardware [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1037.071523] env[61970]: DEBUG nova.virt.hardware [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1037.071523] env[61970]: DEBUG nova.virt.hardware [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1037.071523] env[61970]: DEBUG nova.virt.hardware [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1037.071523] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a397cb2d-836f-4ebc-95f8-9a0a184ae6dc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.088428] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7144eee-4d21-4c30-ae5b-3a0d7484b3bd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.112810] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:cf:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b356db78-99c7-4464-822c-fc7e193f7878', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4e65fd7e-c508-4d66-a854-d26fd273014a', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1037.120645] env[61970]: DEBUG oslo.service.loopingcall [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1037.120985] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1037.121246] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc124c21-958f-4297-8f1d-8188065a6514 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.144711] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1037.144711] env[61970]: value = "task-1356091" [ 1037.144711] env[61970]: _type = "Task" [ 1037.144711] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.156113] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356091, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.206054] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.206342] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.206553] env[61970]: DEBUG nova.network.neutron [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1037.482767] env[61970]: DEBUG nova.network.neutron [-] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.510062] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3ca9cb7d-88c5-4295-a66d-cf0526278637 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.522675] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-075ea0ea-8261-4c2c-87c1-3fce0a98598b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.553297] env[61970]: DEBUG nova.compute.manager [req-9eafe46b-daa4-404a-ae9a-35f73b93ade4 req-5202f602-aa18-4218-bb35-fcc7a3aab29a service nova] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Detach interface failed, port_id=62d22328-b863-41dc-816f-1ce3f7a53eed, reason: Instance d38bde37-994b-4d58-b4b1-5f79b59fd2aa could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1037.655718] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356091, 'name': CreateVM_Task, 'duration_secs': 0.412554} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.655965] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1037.659799] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.659799] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.659799] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1037.659799] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0d6feb1-9f30-47f1-b49f-0c3d633012dc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.663230] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1037.663230] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]527718b5-86b5-fa8f-a4b6-286c68184683" [ 1037.663230] env[61970]: _type = "Task" [ 1037.663230] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.672027] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]527718b5-86b5-fa8f-a4b6-286c68184683, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.984987] env[61970]: INFO nova.compute.manager [-] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Took 1.22 seconds to deallocate network for instance. [ 1038.010315] env[61970]: DEBUG nova.network.neutron [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updating instance_info_cache with network_info: [{"id": "9a1c8b46-0249-4e0b-94de-b0d0062ca69c", "address": "fa:16:3e:93:5e:22", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a1c8b46-02", "ovs_interfaceid": "9a1c8b46-0249-4e0b-94de-b0d0062ca69c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.034599] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Acquiring lock "444939f3-f0d7-4af6-a3fa-40c0a8459839" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.034929] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Lock "444939f3-f0d7-4af6-a3fa-40c0a8459839" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.175919] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]527718b5-86b5-fa8f-a4b6-286c68184683, 'name': SearchDatastore_Task, 'duration_secs': 0.01592} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.176244] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.176480] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1038.176711] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.176861] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.177050] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1038.177319] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2bce6b28-ae12-4491-a8a4-1580fdff81e8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.186396] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1038.186586] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1038.187318] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2cd34d7-0896-4ef7-8d2c-e5e39f1a92f1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.192748] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1038.192748] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5229c3e3-cf11-fe9f-b407-1f21c91195ab" [ 1038.192748] env[61970]: _type = "Task" [ 1038.192748] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.200811] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5229c3e3-cf11-fe9f-b407-1f21c91195ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.333839] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.334206] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.334488] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.334708] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.334925] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.338453] env[61970]: INFO nova.compute.manager [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Terminating instance [ 1038.340211] env[61970]: DEBUG nova.compute.manager [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1038.340441] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1038.341404] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e5bf75-5870-441f-851a-96e2206dd3c6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.350962] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1038.351227] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c31ec178-e4fd-40d9-a0cb-40a3c2e8390c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.359267] env[61970]: DEBUG oslo_vmware.api [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1038.359267] env[61970]: value = "task-1356092" [ 1038.359267] env[61970]: _type = "Task" [ 1038.359267] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.367031] env[61970]: DEBUG oslo_vmware.api [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356092, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.495771] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.496075] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.496295] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.513543] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.526158] env[61970]: INFO nova.scheduler.client.report [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Deleted allocations for instance d38bde37-994b-4d58-b4b1-5f79b59fd2aa [ 1038.537521] env[61970]: DEBUG nova.compute.manager [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1038.704080] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5229c3e3-cf11-fe9f-b407-1f21c91195ab, 'name': SearchDatastore_Task, 'duration_secs': 0.008736} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.704848] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3ec4464-467f-4cda-9f12-f3c05d4f4bd2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.709992] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1038.709992] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f185b8-88c0-1c15-f49e-d7c7f4b625c3" [ 1038.709992] env[61970]: _type = "Task" [ 1038.709992] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.717693] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f185b8-88c0-1c15-f49e-d7c7f4b625c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.869078] env[61970]: DEBUG oslo_vmware.api [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356092, 'name': PowerOffVM_Task, 'duration_secs': 0.203664} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.869380] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1038.869561] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1038.869815] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8dfc907d-9247-4a3a-9036-a8c1467b47c9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.935813] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1038.936103] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1038.936308] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Deleting the datastore file [datastore2] d0ad9689-bdb7-4d68-b19a-3f805ab4612e {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1038.936618] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-60121d4e-d5f0-497c-8c68-cbc9b0b25f19 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.944189] env[61970]: DEBUG oslo_vmware.api [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1038.944189] env[61970]: value = "task-1356094" [ 1038.944189] env[61970]: _type = "Task" [ 1038.944189] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.951838] env[61970]: DEBUG oslo_vmware.api [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356094, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.036243] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f45e0f-c425-425c-a535-92578cb4219f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.039189] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c6baf931-a239-40e5-b694-eb02f373e497 tempest-DeleteServersTestJSON-263811819 tempest-DeleteServersTestJSON-263811819-project-member] Lock "d38bde37-994b-4d58-b4b1-5f79b59fd2aa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.425s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.061445] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e53eef-0e85-4a27-9561-7050a4b41bd8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.070196] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updating instance 'ebb92a5b-f635-4504-a57f-395de514015a' progress to 83 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1039.074910] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.075182] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.076688] env[61970]: INFO nova.compute.claims [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1039.222039] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f185b8-88c0-1c15-f49e-d7c7f4b625c3, 'name': SearchDatastore_Task, 'duration_secs': 0.010264} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.222383] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.222447] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] bc948a41-1a22-46fa-a76a-50b4c4afa4b7/bc948a41-1a22-46fa-a76a-50b4c4afa4b7.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1039.222702] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1fdb9f51-1019-4889-a991-db2fdf4ca014 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.229803] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1039.229803] env[61970]: value = "task-1356095" [ 1039.229803] env[61970]: _type = "Task" [ 1039.229803] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.238100] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356095, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.456765] env[61970]: DEBUG oslo_vmware.api [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356094, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145185} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.457477] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1039.457764] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1039.458026] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1039.458288] env[61970]: INFO nova.compute.manager [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1039.458675] env[61970]: DEBUG oslo.service.loopingcall [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.458959] env[61970]: DEBUG nova.compute.manager [-] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1039.459658] env[61970]: DEBUG nova.network.neutron [-] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1039.506281] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1039.506281] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1039.506281] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Starting heal instance info cache {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1039.576622] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d231e56c-627b-4504-b816-538dcdd3e95a tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updating instance 'ebb92a5b-f635-4504-a57f-395de514015a' progress to 100 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1039.740236] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356095, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488935} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.740537] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] bc948a41-1a22-46fa-a76a-50b4c4afa4b7/bc948a41-1a22-46fa-a76a-50b4c4afa4b7.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1039.740753] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1039.741023] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1b1d74bf-18bd-4820-a5ae-c1f2b6f3fa36 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.766182] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1039.766182] env[61970]: value = "task-1356096" [ 1039.766182] env[61970]: _type = "Task" [ 1039.766182] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.775520] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356096, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.984318] env[61970]: DEBUG nova.compute.manager [req-c3aa91e7-6f9a-4bfa-9287-b01a59928435 req-3645805f-e971-40b4-8a34-9e4beedfe401 service nova] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Received event network-vif-deleted-b83024be-4171-40cd-9d1c-65a2d2ce8b9e {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.984469] env[61970]: INFO nova.compute.manager [req-c3aa91e7-6f9a-4bfa-9287-b01a59928435 req-3645805f-e971-40b4-8a34-9e4beedfe401 service nova] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Neutron deleted interface b83024be-4171-40cd-9d1c-65a2d2ce8b9e; detaching it from the instance and deleting it from the info cache [ 1039.984634] env[61970]: DEBUG nova.network.neutron [req-c3aa91e7-6f9a-4bfa-9287-b01a59928435 req-3645805f-e971-40b4-8a34-9e4beedfe401 service nova] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.039253] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "refresh_cache-0304640a-6c46-4f57-a8b8-5058da2e3de4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.039408] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquired lock "refresh_cache-0304640a-6c46-4f57-a8b8-5058da2e3de4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.039559] env[61970]: DEBUG nova.network.neutron [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Forcefully refreshing network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1040.217787] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3289e1a6-b1f4-4692-a4a5-ece0a7b8b392 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.225762] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb371e2f-3db6-4094-85ad-e1be88ccc339 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.255202] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39891c39-d6a5-41eb-84b4-4814c50bcec6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.262733] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c36cac0f-8e86-482a-b727-6bd2c56d6bb0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.278782] env[61970]: DEBUG nova.compute.provider_tree [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.282597] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356096, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067859} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.283719] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1040.283816] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2032223d-823f-4383-a03b-d31e220be10e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.307468] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] bc948a41-1a22-46fa-a76a-50b4c4afa4b7/bc948a41-1a22-46fa-a76a-50b4c4afa4b7.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1040.308472] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d347e47f-9e95-464e-9b54-bb974ee271c9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.333145] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1040.333145] env[61970]: value = "task-1356098" [ 1040.333145] env[61970]: _type = "Task" [ 1040.333145] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.344652] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356098, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.463965] env[61970]: DEBUG nova.network.neutron [-] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.487220] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f884b1f4-779e-4bb8-89ee-7978dbcdf6d5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.497371] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b0a582f-6691-4d29-9d16-2ca93f05f954 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.525875] env[61970]: DEBUG nova.compute.manager [req-c3aa91e7-6f9a-4bfa-9287-b01a59928435 req-3645805f-e971-40b4-8a34-9e4beedfe401 service nova] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Detach interface failed, port_id=b83024be-4171-40cd-9d1c-65a2d2ce8b9e, reason: Instance d0ad9689-bdb7-4d68-b19a-3f805ab4612e could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1040.785095] env[61970]: DEBUG nova.scheduler.client.report [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1040.843313] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356098, 'name': ReconfigVM_Task, 'duration_secs': 0.305974} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.843709] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Reconfigured VM instance instance-0000005d to attach disk [datastore1] bc948a41-1a22-46fa-a76a-50b4c4afa4b7/bc948a41-1a22-46fa-a76a-50b4c4afa4b7.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1040.844882] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_name': '/dev/sda', 'encryption_secret_uuid': None, 'size': 0, 'encryption_options': None, 'device_type': 'disk', 'disk_bus': None, 'encrypted': False, 'guest_format': None, 'boot_index': 0, 'encryption_format': None, 'image_id': '7ff5af44-3f73-4c84-82e5-a1543db26d03'}], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': False, 'attachment_id': '082ec49d-1478-459d-a658-076688752682', 'boot_index': None, 'device_type': None, 'mount_device': '/dev/sdb', 'disk_bus': None, 'guest_format': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288868', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'name': 'volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bc948a41-1a22-46fa-a76a-50b4c4afa4b7', 'attached_at': '', 'detached_at': '', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'serial': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2'}, 'volume_type': None}], 'swap': None} {{(pid=61970) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1040.845115] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Volume attach. Driver type: vmdk {{(pid=61970) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1040.845317] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288868', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'name': 'volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bc948a41-1a22-46fa-a76a-50b4c4afa4b7', 'attached_at': '', 'detached_at': '', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'serial': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1040.846112] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48f47ef0-9df0-4908-b5e0-35f03cc11907 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.861117] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7c2aa7a-d572-460f-ab60-b1824b293891 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.885289] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2/volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1040.885553] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80186afb-2513-4643-9ceb-9b94192092bc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.904698] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1040.904698] env[61970]: value = "task-1356099" [ 1040.904698] env[61970]: _type = "Task" [ 1040.904698] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.912378] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356099, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.966708] env[61970]: INFO nova.compute.manager [-] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Took 1.51 seconds to deallocate network for instance. [ 1041.267964] env[61970]: DEBUG nova.network.neutron [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Updating instance_info_cache with network_info: [{"id": "2545e2b0-a432-44ae-a287-0b68ceab7031", "address": "fa:16:3e:a6:77:ea", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2545e2b0-a4", "ovs_interfaceid": "2545e2b0-a432-44ae-a287-0b68ceab7031", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.289806] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.214s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.290366] env[61970]: DEBUG nova.compute.manager [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1041.416229] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356099, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.473218] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.473512] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.473806] env[61970]: DEBUG nova.objects.instance [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lazy-loading 'resources' on Instance uuid d0ad9689-bdb7-4d68-b19a-3f805ab4612e {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.771089] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Releasing lock "refresh_cache-0304640a-6c46-4f57-a8b8-5058da2e3de4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.771488] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Updated the network info_cache for instance {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1041.771743] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.771924] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.772158] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.772338] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.772500] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.772659] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.772786] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61970) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1041.772946] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.794608] env[61970]: DEBUG nova.compute.utils [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1041.796024] env[61970]: DEBUG nova.compute.manager [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1041.796142] env[61970]: DEBUG nova.network.neutron [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1041.837915] env[61970]: DEBUG nova.policy [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b650834fb434308aab9d1137e23ef23', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9ed2bc4152164761b653145e7c720b75', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 1041.916833] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356099, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.098694] env[61970]: DEBUG oslo_concurrency.lockutils [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "ebb92a5b-f635-4504-a57f-395de514015a" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.098955] env[61970]: DEBUG oslo_concurrency.lockutils [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "ebb92a5b-f635-4504-a57f-395de514015a" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.099167] env[61970]: DEBUG nova.compute.manager [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Going to confirm migration 4 {{(pid=61970) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1042.100967] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cf3a775-c984-47b8-8959-01df3341de8e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.110297] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c936574-210b-4d10-b55f-6b95f1bfe3eb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.114147] env[61970]: DEBUG nova.network.neutron [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Successfully created port: b8177d93-2145-46f4-b212-1b5bd5c97e41 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1042.147229] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f20300e-5e71-4aa9-a9bd-ffd58c399fbd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.156017] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533d5975-1521-4bbd-b2fb-eb796b8c9902 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.171634] env[61970]: DEBUG nova.compute.provider_tree [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.275968] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.300192] env[61970]: DEBUG nova.compute.manager [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1042.418085] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356099, 'name': ReconfigVM_Task, 'duration_secs': 1.018276} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.418085] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Reconfigured VM instance instance-0000005d to attach disk [datastore1] volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2/volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1042.422330] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb2efec7-dc82-4712-89bf-f47762aea9c8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.437787] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1042.437787] env[61970]: value = "task-1356100" [ 1042.437787] env[61970]: _type = "Task" [ 1042.437787] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.445695] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356100, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.638500] env[61970]: DEBUG oslo_concurrency.lockutils [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.638746] env[61970]: DEBUG oslo_concurrency.lockutils [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.638966] env[61970]: DEBUG nova.network.neutron [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1042.639231] env[61970]: DEBUG nova.objects.instance [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lazy-loading 'info_cache' on Instance uuid ebb92a5b-f635-4504-a57f-395de514015a {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.674236] env[61970]: DEBUG nova.scheduler.client.report [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1042.947932] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356100, 'name': ReconfigVM_Task, 'duration_secs': 0.136003} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.948289] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288868', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'name': 'volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bc948a41-1a22-46fa-a76a-50b4c4afa4b7', 'attached_at': '', 'detached_at': '', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'serial': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1042.948883] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f86a9e38-9811-4c04-a4db-1e6d1250cf7e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.955217] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1042.955217] env[61970]: value = "task-1356101" [ 1042.955217] env[61970]: _type = "Task" [ 1042.955217] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.962839] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356101, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.179214] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.706s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.181714] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.906s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.181714] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.181894] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61970) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1043.182761] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbef15ba-a30b-4a9e-ba31-0c4823f093ac {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.192991] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a8ad228-3973-43e6-9026-332c3169dd15 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.208356] env[61970]: INFO nova.scheduler.client.report [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Deleted allocations for instance d0ad9689-bdb7-4d68-b19a-3f805ab4612e [ 1043.210671] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef12fac-bb95-4b33-8b5a-9c3f6f5f4ee8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.223387] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddee705c-473b-4e1f-a882-4c5064862f45 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.250404] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179847MB free_disk=150GB free_vcpus=48 pci_devices=None {{(pid=61970) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1043.250561] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.250754] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.311918] env[61970]: DEBUG nova.compute.manager [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1043.338809] env[61970]: DEBUG nova.virt.hardware [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1043.339078] env[61970]: DEBUG nova.virt.hardware [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1043.339246] env[61970]: DEBUG nova.virt.hardware [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1043.339431] env[61970]: DEBUG nova.virt.hardware [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1043.339582] env[61970]: DEBUG nova.virt.hardware [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1043.339733] env[61970]: DEBUG nova.virt.hardware [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1043.339966] env[61970]: DEBUG nova.virt.hardware [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1043.340202] env[61970]: DEBUG nova.virt.hardware [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1043.340391] env[61970]: DEBUG nova.virt.hardware [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1043.340562] env[61970]: DEBUG nova.virt.hardware [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1043.340740] env[61970]: DEBUG nova.virt.hardware [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1043.341608] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71db210f-5813-459a-8896-f6452c9e818b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.350668] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8c8e2ff-ddc9-4813-880e-818bc1327825 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.465976] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356101, 'name': Rename_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.625456] env[61970]: DEBUG nova.compute.manager [req-3f2ac2fb-1a34-472c-828c-fe6acd49ba83 req-9e1a5c9b-4a89-4e55-9623-df034022431f service nova] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Received event network-vif-plugged-b8177d93-2145-46f4-b212-1b5bd5c97e41 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1043.625528] env[61970]: DEBUG oslo_concurrency.lockutils [req-3f2ac2fb-1a34-472c-828c-fe6acd49ba83 req-9e1a5c9b-4a89-4e55-9623-df034022431f service nova] Acquiring lock "444939f3-f0d7-4af6-a3fa-40c0a8459839-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.625707] env[61970]: DEBUG oslo_concurrency.lockutils [req-3f2ac2fb-1a34-472c-828c-fe6acd49ba83 req-9e1a5c9b-4a89-4e55-9623-df034022431f service nova] Lock "444939f3-f0d7-4af6-a3fa-40c0a8459839-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.625870] env[61970]: DEBUG oslo_concurrency.lockutils [req-3f2ac2fb-1a34-472c-828c-fe6acd49ba83 req-9e1a5c9b-4a89-4e55-9623-df034022431f service nova] Lock "444939f3-f0d7-4af6-a3fa-40c0a8459839-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.626063] env[61970]: DEBUG nova.compute.manager [req-3f2ac2fb-1a34-472c-828c-fe6acd49ba83 req-9e1a5c9b-4a89-4e55-9623-df034022431f service nova] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] No waiting events found dispatching network-vif-plugged-b8177d93-2145-46f4-b212-1b5bd5c97e41 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1043.626245] env[61970]: WARNING nova.compute.manager [req-3f2ac2fb-1a34-472c-828c-fe6acd49ba83 req-9e1a5c9b-4a89-4e55-9623-df034022431f service nova] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Received unexpected event network-vif-plugged-b8177d93-2145-46f4-b212-1b5bd5c97e41 for instance with vm_state building and task_state spawning. [ 1043.717679] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b017d5fb-bff6-4aed-8f27-0fe383609eaa tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d0ad9689-bdb7-4d68-b19a-3f805ab4612e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.383s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.966866] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356101, 'name': Rename_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.107310] env[61970]: DEBUG nova.network.neutron [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Successfully updated port: b8177d93-2145-46f4-b212-1b5bd5c97e41 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1044.275859] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Applying migration context for instance ebb92a5b-f635-4504-a57f-395de514015a as it has an incoming, in-progress migration e20299e1-a255-41de-9c03-b97e0ba982fd. Migration status is confirming {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1044.275859] env[61970]: INFO nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updating resource usage from migration e20299e1-a255-41de-9c03-b97e0ba982fd [ 1044.289713] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 0304640a-6c46-4f57-a8b8-5058da2e3de4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.290067] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance bc948a41-1a22-46fa-a76a-50b4c4afa4b7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.290286] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 7f59000a-94d9-45b6-aa7a-300d95793615 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.290517] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 31f9d37b-f4ff-4a8b-9477-8253f8305020 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.290775] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Migration e20299e1-a255-41de-9c03-b97e0ba982fd is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1044.290958] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance ebb92a5b-f635-4504-a57f-395de514015a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.291292] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 444939f3-f0d7-4af6-a3fa-40c0a8459839 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1044.291479] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1044.291733] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1920MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1044.339666] env[61970]: DEBUG nova.network.neutron [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updating instance_info_cache with network_info: [{"id": "9a1c8b46-0249-4e0b-94de-b0d0062ca69c", "address": "fa:16:3e:93:5e:22", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a1c8b46-02", "ovs_interfaceid": "9a1c8b46-0249-4e0b-94de-b0d0062ca69c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.404363] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-449acbe3-b3ed-4fcc-9d36-5a8375cdc3d6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.413326] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a4ddd9c-baf9-402d-ac8f-59943fb2f097 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.445213] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b49c1d2-e089-4e61-baf7-194e76e3e9a9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.454616] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3776c851-e89a-4c3c-b3f2-311108af2ff4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.480812] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1044.482320] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356101, 'name': Rename_Task, 'duration_secs': 1.151082} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.482799] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1044.483143] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-02814920-ec40-4f2c-9f4f-a6bde69af9f1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.492477] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1044.492477] env[61970]: value = "task-1356102" [ 1044.492477] env[61970]: _type = "Task" [ 1044.492477] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.504357] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356102, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.610185] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Acquiring lock "refresh_cache-444939f3-f0d7-4af6-a3fa-40c0a8459839" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.610537] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Acquired lock "refresh_cache-444939f3-f0d7-4af6-a3fa-40c0a8459839" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.610713] env[61970]: DEBUG nova.network.neutron [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1044.843667] env[61970]: DEBUG oslo_concurrency.lockutils [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.844068] env[61970]: DEBUG nova.objects.instance [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lazy-loading 'migration_context' on Instance uuid ebb92a5b-f635-4504-a57f-395de514015a {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.984354] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1045.006043] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356102, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.178188] env[61970]: DEBUG nova.network.neutron [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1045.182882] env[61970]: DEBUG oslo_concurrency.lockutils [None req-868c4b08-3469-49a9-b817-247228556410 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "7f59000a-94d9-45b6-aa7a-300d95793615" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.183345] env[61970]: DEBUG oslo_concurrency.lockutils [None req-868c4b08-3469-49a9-b817-247228556410 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "7f59000a-94d9-45b6-aa7a-300d95793615" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.347464] env[61970]: DEBUG nova.objects.base [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1045.348493] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af992bc5-395b-417d-b5a7-07572b741514 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.378350] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48049f17-80f7-44dd-8960-574e4b9410d4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.385353] env[61970]: DEBUG oslo_vmware.api [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1045.385353] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52bcb67d-c9e7-0c6d-c574-f01a872e2f68" [ 1045.385353] env[61970]: _type = "Task" [ 1045.385353] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.396010] env[61970]: DEBUG oslo_vmware.api [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52bcb67d-c9e7-0c6d-c574-f01a872e2f68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.437682] env[61970]: DEBUG nova.network.neutron [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Updating instance_info_cache with network_info: [{"id": "b8177d93-2145-46f4-b212-1b5bd5c97e41", "address": "fa:16:3e:ff:7a:b5", "network": {"id": "90ea934c-7dce-46f4-ba43-ffc4d271a722", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-43559245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ed2bc4152164761b653145e7c720b75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8177d93-21", "ovs_interfaceid": "b8177d93-2145-46f4-b212-1b5bd5c97e41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.492025] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1045.492175] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.241s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.509271] env[61970]: DEBUG oslo_vmware.api [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356102, 'name': PowerOnVM_Task, 'duration_secs': 0.530655} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.509271] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1045.509409] env[61970]: DEBUG nova.compute.manager [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1045.510653] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d13ae11-e8ef-46ab-9883-79aabcd33871 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.597185] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.597483] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.649531] env[61970]: DEBUG nova.compute.manager [req-b13e2c96-47c1-4ea9-aab9-df9d21adac75 req-90338991-a3ac-4745-8ea5-1023fd7e8d8c service nova] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Received event network-changed-b8177d93-2145-46f4-b212-1b5bd5c97e41 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.649749] env[61970]: DEBUG nova.compute.manager [req-b13e2c96-47c1-4ea9-aab9-df9d21adac75 req-90338991-a3ac-4745-8ea5-1023fd7e8d8c service nova] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Refreshing instance network info cache due to event network-changed-b8177d93-2145-46f4-b212-1b5bd5c97e41. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1045.650034] env[61970]: DEBUG oslo_concurrency.lockutils [req-b13e2c96-47c1-4ea9-aab9-df9d21adac75 req-90338991-a3ac-4745-8ea5-1023fd7e8d8c service nova] Acquiring lock "refresh_cache-444939f3-f0d7-4af6-a3fa-40c0a8459839" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.687141] env[61970]: INFO nova.compute.manager [None req-868c4b08-3469-49a9-b817-247228556410 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Detaching volume a78067a3-226b-4aec-bb5b-6fb37d5bfca3 [ 1045.722978] env[61970]: INFO nova.virt.block_device [None req-868c4b08-3469-49a9-b817-247228556410 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Attempting to driver detach volume a78067a3-226b-4aec-bb5b-6fb37d5bfca3 from mountpoint /dev/sdb [ 1045.723271] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-868c4b08-3469-49a9-b817-247228556410 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Volume detach. Driver type: vmdk {{(pid=61970) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1045.723561] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-868c4b08-3469-49a9-b817-247228556410 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288853', 'volume_id': 'a78067a3-226b-4aec-bb5b-6fb37d5bfca3', 'name': 'volume-a78067a3-226b-4aec-bb5b-6fb37d5bfca3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '7f59000a-94d9-45b6-aa7a-300d95793615', 'attached_at': '', 'detached_at': '', 'volume_id': 'a78067a3-226b-4aec-bb5b-6fb37d5bfca3', 'serial': 'a78067a3-226b-4aec-bb5b-6fb37d5bfca3'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1045.724458] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c2ac923-379f-4bbc-854c-9bc1f0fabd7a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.748910] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f62a62f8-f303-4c0f-ba82-b3b65a4b7e3e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.760550] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca8fa6b9-8c5f-478f-995c-31e7db307f77 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.780979] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c7fa98c-b94a-4d8a-9170-26a44dc0e4e1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.796736] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-868c4b08-3469-49a9-b817-247228556410 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] The volume has not been displaced from its original location: [datastore2] volume-a78067a3-226b-4aec-bb5b-6fb37d5bfca3/volume-a78067a3-226b-4aec-bb5b-6fb37d5bfca3.vmdk. No consolidation needed. {{(pid=61970) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1045.802139] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-868c4b08-3469-49a9-b817-247228556410 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Reconfiguring VM instance instance-00000050 to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1045.802454] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2dc0a110-6c1f-4b41-ab24-e4a72b54f8a7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.821516] env[61970]: DEBUG oslo_vmware.api [None req-868c4b08-3469-49a9-b817-247228556410 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1045.821516] env[61970]: value = "task-1356103" [ 1045.821516] env[61970]: _type = "Task" [ 1045.821516] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.829801] env[61970]: DEBUG oslo_vmware.api [None req-868c4b08-3469-49a9-b817-247228556410 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356103, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.900092] env[61970]: DEBUG oslo_vmware.api [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52bcb67d-c9e7-0c6d-c574-f01a872e2f68, 'name': SearchDatastore_Task, 'duration_secs': 0.009027} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.900641] env[61970]: DEBUG oslo_concurrency.lockutils [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.901059] env[61970]: DEBUG oslo_concurrency.lockutils [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.940929] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Releasing lock "refresh_cache-444939f3-f0d7-4af6-a3fa-40c0a8459839" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.941315] env[61970]: DEBUG nova.compute.manager [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Instance network_info: |[{"id": "b8177d93-2145-46f4-b212-1b5bd5c97e41", "address": "fa:16:3e:ff:7a:b5", "network": {"id": "90ea934c-7dce-46f4-ba43-ffc4d271a722", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-43559245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ed2bc4152164761b653145e7c720b75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8177d93-21", "ovs_interfaceid": "b8177d93-2145-46f4-b212-1b5bd5c97e41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1045.941595] env[61970]: DEBUG oslo_concurrency.lockutils [req-b13e2c96-47c1-4ea9-aab9-df9d21adac75 req-90338991-a3ac-4745-8ea5-1023fd7e8d8c service nova] Acquired lock "refresh_cache-444939f3-f0d7-4af6-a3fa-40c0a8459839" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.941780] env[61970]: DEBUG nova.network.neutron [req-b13e2c96-47c1-4ea9-aab9-df9d21adac75 req-90338991-a3ac-4745-8ea5-1023fd7e8d8c service nova] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Refreshing network info cache for port b8177d93-2145-46f4-b212-1b5bd5c97e41 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1045.943099] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ff:7a:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4df917f7-847a-4c0e-b0e3-69a52e4a1554', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b8177d93-2145-46f4-b212-1b5bd5c97e41', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1045.951244] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Creating folder: Project (9ed2bc4152164761b653145e7c720b75). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1045.952400] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-84fa15f8-0364-4e06-a2b4-6734c30766cc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.966972] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Created folder: Project (9ed2bc4152164761b653145e7c720b75) in parent group-v288740. [ 1045.967227] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Creating folder: Instances. Parent ref: group-v288870. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1045.967539] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d50778fe-195a-4822-95bc-0c00a67e3547 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.978738] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Created folder: Instances in parent group-v288870. [ 1045.979084] env[61970]: DEBUG oslo.service.loopingcall [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1045.979765] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1045.979765] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ad73f43-d254-4238-8e92-99f1e971e449 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.006072] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1046.006072] env[61970]: value = "task-1356106" [ 1046.006072] env[61970]: _type = "Task" [ 1046.006072] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.014649] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356106, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.032545] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.099806] env[61970]: DEBUG nova.compute.manager [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1046.334202] env[61970]: DEBUG oslo_vmware.api [None req-868c4b08-3469-49a9-b817-247228556410 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356103, 'name': ReconfigVM_Task, 'duration_secs': 0.239271} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.334589] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-868c4b08-3469-49a9-b817-247228556410 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Reconfigured VM instance instance-00000050 to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1046.340255] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db32a7a4-0829-497e-8171-2bcf99dffd26 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.357453] env[61970]: DEBUG oslo_vmware.api [None req-868c4b08-3469-49a9-b817-247228556410 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1046.357453] env[61970]: value = "task-1356107" [ 1046.357453] env[61970]: _type = "Task" [ 1046.357453] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.368678] env[61970]: DEBUG oslo_vmware.api [None req-868c4b08-3469-49a9-b817-247228556410 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356107, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.517384] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356106, 'name': CreateVM_Task, 'duration_secs': 0.36145} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.517604] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1046.520322] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.520322] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.520322] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1046.521398] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a281607-82f8-4150-88b6-2aa3f72b2486 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.527198] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Waiting for the task: (returnval){ [ 1046.527198] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e270b1-b11e-4fc0-c3ad-247656b831a8" [ 1046.527198] env[61970]: _type = "Task" [ 1046.527198] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.535661] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e270b1-b11e-4fc0-c3ad-247656b831a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.559396] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b36631f-63b8-4e56-9a88-1f2c9e040fc5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.568962] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21358dcb-cc09-43ff-a82a-a84cfeece724 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.602600] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d108e514-4dea-4ff4-8579-cc9c4b73e357 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.614926] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca6f9fe-44d3-419a-a684-a00e6d2dfebd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.629228] env[61970]: DEBUG nova.compute.provider_tree [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.631335] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.826587] env[61970]: DEBUG nova.network.neutron [req-b13e2c96-47c1-4ea9-aab9-df9d21adac75 req-90338991-a3ac-4745-8ea5-1023fd7e8d8c service nova] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Updated VIF entry in instance network info cache for port b8177d93-2145-46f4-b212-1b5bd5c97e41. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1046.826973] env[61970]: DEBUG nova.network.neutron [req-b13e2c96-47c1-4ea9-aab9-df9d21adac75 req-90338991-a3ac-4745-8ea5-1023fd7e8d8c service nova] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Updating instance_info_cache with network_info: [{"id": "b8177d93-2145-46f4-b212-1b5bd5c97e41", "address": "fa:16:3e:ff:7a:b5", "network": {"id": "90ea934c-7dce-46f4-ba43-ffc4d271a722", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-43559245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ed2bc4152164761b653145e7c720b75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8177d93-21", "ovs_interfaceid": "b8177d93-2145-46f4-b212-1b5bd5c97e41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.867988] env[61970]: DEBUG oslo_vmware.api [None req-868c4b08-3469-49a9-b817-247228556410 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356107, 'name': ReconfigVM_Task, 'duration_secs': 0.147556} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.868318] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-868c4b08-3469-49a9-b817-247228556410 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288853', 'volume_id': 'a78067a3-226b-4aec-bb5b-6fb37d5bfca3', 'name': 'volume-a78067a3-226b-4aec-bb5b-6fb37d5bfca3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '7f59000a-94d9-45b6-aa7a-300d95793615', 'attached_at': '', 'detached_at': '', 'volume_id': 'a78067a3-226b-4aec-bb5b-6fb37d5bfca3', 'serial': 'a78067a3-226b-4aec-bb5b-6fb37d5bfca3'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1047.038105] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e270b1-b11e-4fc0-c3ad-247656b831a8, 'name': SearchDatastore_Task, 'duration_secs': 0.038859} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.038418] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.038655] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1047.038896] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.039064] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.039254] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1047.039514] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ae9e60ee-2db5-4278-8361-f27beb6b02c0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.055901] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1047.056191] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1047.057267] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9ea1b16-0574-4d2e-b766-a3196dc1f37d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.063191] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Waiting for the task: (returnval){ [ 1047.063191] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5274d6f7-8ce7-1d82-d761-96b04df172fb" [ 1047.063191] env[61970]: _type = "Task" [ 1047.063191] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.072478] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5274d6f7-8ce7-1d82-d761-96b04df172fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.133868] env[61970]: DEBUG nova.scheduler.client.report [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1047.330023] env[61970]: DEBUG oslo_concurrency.lockutils [req-b13e2c96-47c1-4ea9-aab9-df9d21adac75 req-90338991-a3ac-4745-8ea5-1023fd7e8d8c service nova] Releasing lock "refresh_cache-444939f3-f0d7-4af6-a3fa-40c0a8459839" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.415733] env[61970]: DEBUG nova.objects.instance [None req-868c4b08-3469-49a9-b817-247228556410 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lazy-loading 'flavor' on Instance uuid 7f59000a-94d9-45b6-aa7a-300d95793615 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.578158] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5274d6f7-8ce7-1d82-d761-96b04df172fb, 'name': SearchDatastore_Task, 'duration_secs': 0.022942} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.578158] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ceccf9c-fc38-45e0-ba6b-e47e2eedc6fd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.585085] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Waiting for the task: (returnval){ [ 1047.585085] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]522b1fa7-e0d2-75ef-69f8-1bbf8fd2a27c" [ 1047.585085] env[61970]: _type = "Task" [ 1047.585085] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.595348] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]522b1fa7-e0d2-75ef-69f8-1bbf8fd2a27c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.696524] env[61970]: DEBUG oslo_concurrency.lockutils [None req-04126f97-abb7-4f6f-ba70-6ea1c2d421d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.696801] env[61970]: DEBUG oslo_concurrency.lockutils [None req-04126f97-abb7-4f6f-ba70-6ea1c2d421d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.095906] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]522b1fa7-e0d2-75ef-69f8-1bbf8fd2a27c, 'name': SearchDatastore_Task, 'duration_secs': 0.010928} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.096050] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.096314] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 444939f3-f0d7-4af6-a3fa-40c0a8459839/444939f3-f0d7-4af6-a3fa-40c0a8459839.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1048.096573] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c6ef8824-dd40-4579-b66b-d76c67fbe68e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.104236] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Waiting for the task: (returnval){ [ 1048.104236] env[61970]: value = "task-1356108" [ 1048.104236] env[61970]: _type = "Task" [ 1048.104236] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.112304] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': task-1356108, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.144830] env[61970]: DEBUG oslo_concurrency.lockutils [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.244s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.145073] env[61970]: DEBUG nova.compute.manager [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=61970) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 1048.149123] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.117s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.149331] env[61970]: DEBUG nova.objects.instance [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61970) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1048.200443] env[61970]: INFO nova.compute.manager [None req-04126f97-abb7-4f6f-ba70-6ea1c2d421d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Detaching volume 8724c3ef-36fc-44b6-a9e3-dfee76f52fa2 [ 1048.246454] env[61970]: INFO nova.virt.block_device [None req-04126f97-abb7-4f6f-ba70-6ea1c2d421d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Attempting to driver detach volume 8724c3ef-36fc-44b6-a9e3-dfee76f52fa2 from mountpoint /dev/sdb [ 1048.246914] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-04126f97-abb7-4f6f-ba70-6ea1c2d421d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Volume detach. Driver type: vmdk {{(pid=61970) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1048.246914] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-04126f97-abb7-4f6f-ba70-6ea1c2d421d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288868', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'name': 'volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bc948a41-1a22-46fa-a76a-50b4c4afa4b7', 'attached_at': '', 'detached_at': '', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'serial': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1048.247804] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90c7456-1a8c-49a0-b624-c656836683b3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.272393] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d14af8b-ebf1-4a30-8895-9e273b453880 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.281131] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4cb94bf-1dfd-4dbe-be44-081683cc2573 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.302794] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-181bf5f2-ab27-4043-9dd4-0ed81fc1445f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.322306] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-04126f97-abb7-4f6f-ba70-6ea1c2d421d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] The volume has not been displaced from its original location: [datastore1] volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2/volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2.vmdk. No consolidation needed. {{(pid=61970) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1048.328874] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-04126f97-abb7-4f6f-ba70-6ea1c2d421d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Reconfiguring VM instance instance-0000005d to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1048.329290] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd1db841-cc59-4886-b503-f25767b4013a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.350055] env[61970]: DEBUG oslo_vmware.api [None req-04126f97-abb7-4f6f-ba70-6ea1c2d421d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1048.350055] env[61970]: value = "task-1356109" [ 1048.350055] env[61970]: _type = "Task" [ 1048.350055] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.358066] env[61970]: DEBUG oslo_vmware.api [None req-04126f97-abb7-4f6f-ba70-6ea1c2d421d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356109, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.424554] env[61970]: DEBUG oslo_concurrency.lockutils [None req-868c4b08-3469-49a9-b817-247228556410 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "7f59000a-94d9-45b6-aa7a-300d95793615" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.240s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.615696] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': task-1356108, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.724043] env[61970]: INFO nova.scheduler.client.report [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Deleted allocation for migration e20299e1-a255-41de-9c03-b97e0ba982fd [ 1048.859667] env[61970]: DEBUG oslo_vmware.api [None req-04126f97-abb7-4f6f-ba70-6ea1c2d421d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356109, 'name': ReconfigVM_Task, 'duration_secs': 0.366096} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.859972] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-04126f97-abb7-4f6f-ba70-6ea1c2d421d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Reconfigured VM instance instance-0000005d to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1048.864760] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23e552f8-0c76-4412-9c9c-1001cb58e380 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.881718] env[61970]: DEBUG oslo_vmware.api [None req-04126f97-abb7-4f6f-ba70-6ea1c2d421d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1048.881718] env[61970]: value = "task-1356110" [ 1048.881718] env[61970]: _type = "Task" [ 1048.881718] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.890301] env[61970]: DEBUG oslo_vmware.api [None req-04126f97-abb7-4f6f-ba70-6ea1c2d421d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356110, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.060053] env[61970]: DEBUG oslo_concurrency.lockutils [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Acquiring lock "ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.060340] env[61970]: DEBUG oslo_concurrency.lockutils [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Lock "ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.115339] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': task-1356108, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.753243} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.115600] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 444939f3-f0d7-4af6-a3fa-40c0a8459839/444939f3-f0d7-4af6-a3fa-40c0a8459839.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1049.115818] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1049.116091] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9d0bf960-a007-426e-9142-cab150ec8e29 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.124034] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Waiting for the task: (returnval){ [ 1049.124034] env[61970]: value = "task-1356111" [ 1049.124034] env[61970]: _type = "Task" [ 1049.124034] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.132164] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': task-1356111, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.164070] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c379ce8d-ae6e-4e18-9c49-8c79c16c1962 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.165272] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.534s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.166768] env[61970]: INFO nova.compute.claims [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1049.233738] env[61970]: DEBUG oslo_concurrency.lockutils [None req-237a6dfa-4d1d-404c-839b-7d05a7f4b76b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "ebb92a5b-f635-4504-a57f-395de514015a" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.135s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.392792] env[61970]: DEBUG oslo_vmware.api [None req-04126f97-abb7-4f6f-ba70-6ea1c2d421d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356110, 'name': ReconfigVM_Task, 'duration_secs': 0.186294} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.393168] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-04126f97-abb7-4f6f-ba70-6ea1c2d421d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288868', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'name': 'volume-8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bc948a41-1a22-46fa-a76a-50b4c4afa4b7', 'attached_at': '', 'detached_at': '', 'volume_id': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2', 'serial': '8724c3ef-36fc-44b6-a9e3-dfee76f52fa2'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1049.439241] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "7f59000a-94d9-45b6-aa7a-300d95793615" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.439525] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "7f59000a-94d9-45b6-aa7a-300d95793615" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.439942] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "7f59000a-94d9-45b6-aa7a-300d95793615-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.440213] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "7f59000a-94d9-45b6-aa7a-300d95793615-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.440428] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "7f59000a-94d9-45b6-aa7a-300d95793615-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.442627] env[61970]: INFO nova.compute.manager [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Terminating instance [ 1049.444440] env[61970]: DEBUG nova.compute.manager [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1049.444635] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1049.445468] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a10a6344-4af4-4761-bc32-7b20ddfb45d0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.454010] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1049.454257] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-17f13c5b-6ae1-4aa4-9f7f-87412bb3d46d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.460822] env[61970]: DEBUG oslo_vmware.api [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1049.460822] env[61970]: value = "task-1356112" [ 1049.460822] env[61970]: _type = "Task" [ 1049.460822] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.469311] env[61970]: DEBUG nova.objects.instance [None req-a13e67e5-707d-4b88-96d6-a24d4e48bd21 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lazy-loading 'flavor' on Instance uuid ebb92a5b-f635-4504-a57f-395de514015a {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.470679] env[61970]: DEBUG oslo_vmware.api [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356112, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.563074] env[61970]: DEBUG nova.compute.manager [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1049.634097] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': task-1356111, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.214745} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.634783] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1049.635875] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d4522e0-4587-4dff-b05c-76c82ad68ae6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.657688] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 444939f3-f0d7-4af6-a3fa-40c0a8459839/444939f3-f0d7-4af6-a3fa-40c0a8459839.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1049.657996] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bdaa31bd-94f1-4ae3-9bf9-5784830138a1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.680087] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Waiting for the task: (returnval){ [ 1049.680087] env[61970]: value = "task-1356113" [ 1049.680087] env[61970]: _type = "Task" [ 1049.680087] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.689684] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': task-1356113, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.938113] env[61970]: DEBUG nova.objects.instance [None req-04126f97-abb7-4f6f-ba70-6ea1c2d421d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lazy-loading 'flavor' on Instance uuid bc948a41-1a22-46fa-a76a-50b4c4afa4b7 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.973121] env[61970]: DEBUG oslo_vmware.api [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356112, 'name': PowerOffVM_Task, 'duration_secs': 0.199479} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.973557] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a13e67e5-707d-4b88-96d6-a24d4e48bd21 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.973715] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a13e67e5-707d-4b88-96d6-a24d4e48bd21 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.973889] env[61970]: DEBUG nova.network.neutron [None req-a13e67e5-707d-4b88-96d6-a24d4e48bd21 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1049.974084] env[61970]: DEBUG nova.objects.instance [None req-a13e67e5-707d-4b88-96d6-a24d4e48bd21 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lazy-loading 'info_cache' on Instance uuid ebb92a5b-f635-4504-a57f-395de514015a {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.975732] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1049.975732] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1049.975732] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-79f50c0d-d973-46c0-9df5-86a5a25635df {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.041837] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1050.042119] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1050.042321] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Deleting the datastore file [datastore2] 7f59000a-94d9-45b6-aa7a-300d95793615 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1050.042595] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-89eb0885-0179-4944-a040-b7477d489c67 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.048970] env[61970]: DEBUG oslo_vmware.api [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1050.048970] env[61970]: value = "task-1356115" [ 1050.048970] env[61970]: _type = "Task" [ 1050.048970] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.057303] env[61970]: DEBUG oslo_vmware.api [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356115, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.085692] env[61970]: DEBUG oslo_concurrency.lockutils [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.194152] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': task-1356113, 'name': ReconfigVM_Task, 'duration_secs': 0.257803} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.194434] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 444939f3-f0d7-4af6-a3fa-40c0a8459839/444939f3-f0d7-4af6-a3fa-40c0a8459839.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.197429] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-59d4551e-2e54-405e-a382-8e9e1e9d6e6f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.205346] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Waiting for the task: (returnval){ [ 1050.205346] env[61970]: value = "task-1356116" [ 1050.205346] env[61970]: _type = "Task" [ 1050.205346] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.214812] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': task-1356116, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.296375] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a17c791-f386-4f1b-8f1d-ed7c4998b25f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.305215] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba5d9a3-da5f-48b2-97b1-bbc5e0cafee5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.337124] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9661dab7-56c1-4e02-8f1a-45fb71ec30cf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.346034] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f9138d5-c7e4-41fb-a568-6aa1f6bede05 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.359836] env[61970]: DEBUG nova.compute.provider_tree [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1050.477248] env[61970]: DEBUG nova.objects.base [None req-a13e67e5-707d-4b88-96d6-a24d4e48bd21 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1050.561057] env[61970]: DEBUG oslo_vmware.api [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356115, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.315334} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.561057] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1050.561057] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1050.561236] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1050.561285] env[61970]: INFO nova.compute.manager [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1050.561529] env[61970]: DEBUG oslo.service.loopingcall [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.561765] env[61970]: DEBUG nova.compute.manager [-] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1050.561847] env[61970]: DEBUG nova.network.neutron [-] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1050.716854] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': task-1356116, 'name': Rename_Task, 'duration_secs': 0.158331} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.717148] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1050.717412] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f2c469b-bc17-4d87-9ed1-310cacb22a75 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.724749] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Waiting for the task: (returnval){ [ 1050.724749] env[61970]: value = "task-1356117" [ 1050.724749] env[61970]: _type = "Task" [ 1050.724749] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.734169] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': task-1356117, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.863054] env[61970]: DEBUG nova.scheduler.client.report [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1050.946792] env[61970]: DEBUG oslo_concurrency.lockutils [None req-04126f97-abb7-4f6f-ba70-6ea1c2d421d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.250s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.015588] env[61970]: DEBUG nova.compute.manager [req-e0fe58b8-48d9-46b4-aff2-1dfae5a2c2ea req-13afc419-1559-4e0c-8134-74fffa0396e2 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Received event network-vif-deleted-5c475b00-f4ed-4e09-916c-6f3ca7845784 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1051.015787] env[61970]: INFO nova.compute.manager [req-e0fe58b8-48d9-46b4-aff2-1dfae5a2c2ea req-13afc419-1559-4e0c-8134-74fffa0396e2 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Neutron deleted interface 5c475b00-f4ed-4e09-916c-6f3ca7845784; detaching it from the instance and deleting it from the info cache [ 1051.015958] env[61970]: DEBUG nova.network.neutron [req-e0fe58b8-48d9-46b4-aff2-1dfae5a2c2ea req-13afc419-1559-4e0c-8134-74fffa0396e2 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.218372] env[61970]: DEBUG nova.network.neutron [None req-a13e67e5-707d-4b88-96d6-a24d4e48bd21 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updating instance_info_cache with network_info: [{"id": "9a1c8b46-0249-4e0b-94de-b0d0062ca69c", "address": "fa:16:3e:93:5e:22", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a1c8b46-02", "ovs_interfaceid": "9a1c8b46-0249-4e0b-94de-b0d0062ca69c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.237349] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': task-1356117, 'name': PowerOnVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.368311] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.203s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.369035] env[61970]: DEBUG nova.compute.manager [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1051.373627] env[61970]: DEBUG oslo_concurrency.lockutils [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.288s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.376345] env[61970]: INFO nova.compute.claims [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1051.485052] env[61970]: DEBUG nova.network.neutron [-] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.522035] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cdf34c04-fd01-4190-804a-84b7900b7be8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.532167] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45db284-6f36-42e5-b754-8705277b2167 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.562154] env[61970]: DEBUG nova.compute.manager [req-e0fe58b8-48d9-46b4-aff2-1dfae5a2c2ea req-13afc419-1559-4e0c-8134-74fffa0396e2 service nova] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Detach interface failed, port_id=5c475b00-f4ed-4e09-916c-6f3ca7845784, reason: Instance 7f59000a-94d9-45b6-aa7a-300d95793615 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1051.721887] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a13e67e5-707d-4b88-96d6-a24d4e48bd21 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "refresh_cache-ebb92a5b-f635-4504-a57f-395de514015a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.739670] env[61970]: DEBUG oslo_vmware.api [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': task-1356117, 'name': PowerOnVM_Task, 'duration_secs': 0.692318} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.739670] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1051.739670] env[61970]: INFO nova.compute.manager [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Took 8.43 seconds to spawn the instance on the hypervisor. [ 1051.739670] env[61970]: DEBUG nova.compute.manager [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1051.740469] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-139eef21-35fa-40d3-b017-8d7691898450 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.882515] env[61970]: DEBUG nova.compute.utils [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1051.885657] env[61970]: DEBUG nova.compute.manager [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1051.885823] env[61970]: DEBUG nova.network.neutron [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1051.931387] env[61970]: DEBUG nova.policy [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb2ab679e39540f09ce1fa95971677f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a44c05a894394d69b8b4fd36c7b2532a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 1051.962710] env[61970]: DEBUG oslo_concurrency.lockutils [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.963301] env[61970]: DEBUG oslo_concurrency.lockutils [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.963301] env[61970]: DEBUG oslo_concurrency.lockutils [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.963468] env[61970]: DEBUG oslo_concurrency.lockutils [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.963609] env[61970]: DEBUG oslo_concurrency.lockutils [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.965809] env[61970]: INFO nova.compute.manager [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Terminating instance [ 1051.967586] env[61970]: DEBUG nova.compute.manager [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1051.967784] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1051.968629] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27bd8a30-7e02-432b-8a72-8233b35a2d86 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.977221] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1051.977422] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-774a1593-5511-4a05-af91-b7dd687a3233 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.983935] env[61970]: DEBUG oslo_vmware.api [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1051.983935] env[61970]: value = "task-1356118" [ 1051.983935] env[61970]: _type = "Task" [ 1051.983935] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.987086] env[61970]: INFO nova.compute.manager [-] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Took 1.43 seconds to deallocate network for instance. [ 1051.996531] env[61970]: DEBUG oslo_vmware.api [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356118, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.205145] env[61970]: DEBUG nova.network.neutron [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Successfully created port: a7b3f775-a79c-488c-bec7-e8a2b425c91f {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1052.225449] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-a13e67e5-707d-4b88-96d6-a24d4e48bd21 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1052.225729] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6cde9a8f-efb3-40e8-a54a-e3875c19e4ba {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.233997] env[61970]: DEBUG oslo_vmware.api [None req-a13e67e5-707d-4b88-96d6-a24d4e48bd21 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1052.233997] env[61970]: value = "task-1356119" [ 1052.233997] env[61970]: _type = "Task" [ 1052.233997] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.242086] env[61970]: DEBUG oslo_vmware.api [None req-a13e67e5-707d-4b88-96d6-a24d4e48bd21 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356119, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.261727] env[61970]: INFO nova.compute.manager [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Took 13.22 seconds to build instance. [ 1052.386586] env[61970]: DEBUG nova.compute.manager [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1052.499140] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.499465] env[61970]: DEBUG oslo_vmware.api [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356118, 'name': PowerOffVM_Task, 'duration_secs': 0.213061} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.502498] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1052.502702] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1052.503999] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-81785b7e-6191-4c2c-88dc-a66699418e06 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.530586] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f34bea8-476a-43fd-876e-44202299daf7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.539722] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e3b7706-93f3-4dfe-b0ce-04541d7155d6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.575214] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf22289-df6d-4e8d-aa73-11c14180b281 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.579425] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1052.579697] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1052.579995] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleting the datastore file [datastore1] bc948a41-1a22-46fa-a76a-50b4c4afa4b7 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1052.580933] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-efbc2dbe-ebdb-4eb1-8cf1-ff0ab14955c2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.586489] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c275d6-7a07-49f9-a977-2ffc62fba77c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.592225] env[61970]: DEBUG oslo_vmware.api [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1052.592225] env[61970]: value = "task-1356121" [ 1052.592225] env[61970]: _type = "Task" [ 1052.592225] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.606571] env[61970]: DEBUG nova.compute.provider_tree [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1052.611959] env[61970]: DEBUG oslo_vmware.api [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356121, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.744177] env[61970]: DEBUG oslo_vmware.api [None req-a13e67e5-707d-4b88-96d6-a24d4e48bd21 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356119, 'name': PowerOnVM_Task, 'duration_secs': 0.423905} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.744989] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-a13e67e5-707d-4b88-96d6-a24d4e48bd21 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1052.745380] env[61970]: DEBUG nova.compute.manager [None req-a13e67e5-707d-4b88-96d6-a24d4e48bd21 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1052.746337] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18bbadae-8403-4dbc-9b1e-dc583bc4ae2f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.765291] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4e2a5402-e477-451e-b9d0-db5c934e5237 tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Lock "444939f3-f0d7-4af6-a3fa-40c0a8459839" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.729s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.101794] env[61970]: DEBUG oslo_vmware.api [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356121, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171296} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.102033] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1053.102228] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1053.102413] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1053.102593] env[61970]: INFO nova.compute.manager [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1053.102875] env[61970]: DEBUG oslo.service.loopingcall [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1053.103086] env[61970]: DEBUG nova.compute.manager [-] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1053.103183] env[61970]: DEBUG nova.network.neutron [-] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1053.108554] env[61970]: DEBUG nova.scheduler.client.report [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1053.380601] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Acquiring lock "444939f3-f0d7-4af6-a3fa-40c0a8459839" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.380695] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Lock "444939f3-f0d7-4af6-a3fa-40c0a8459839" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.380883] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Acquiring lock "444939f3-f0d7-4af6-a3fa-40c0a8459839-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.381071] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Lock "444939f3-f0d7-4af6-a3fa-40c0a8459839-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.381254] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Lock "444939f3-f0d7-4af6-a3fa-40c0a8459839-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.384528] env[61970]: INFO nova.compute.manager [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Terminating instance [ 1053.385377] env[61970]: DEBUG nova.compute.manager [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1053.385593] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1053.386462] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c22892f9-e0b8-4424-973f-f925c000a1bc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.395030] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1053.395030] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5743a6e1-440d-46b4-8f5b-a78111d1a37d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.398139] env[61970]: DEBUG nova.compute.manager [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1053.402802] env[61970]: DEBUG oslo_vmware.api [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Waiting for the task: (returnval){ [ 1053.402802] env[61970]: value = "task-1356122" [ 1053.402802] env[61970]: _type = "Task" [ 1053.402802] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.411954] env[61970]: DEBUG oslo_vmware.api [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': task-1356122, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.434671] env[61970]: DEBUG nova.virt.hardware [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1053.435170] env[61970]: DEBUG nova.virt.hardware [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1053.435170] env[61970]: DEBUG nova.virt.hardware [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1053.435304] env[61970]: DEBUG nova.virt.hardware [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1053.435418] env[61970]: DEBUG nova.virt.hardware [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1053.435572] env[61970]: DEBUG nova.virt.hardware [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1053.435785] env[61970]: DEBUG nova.virt.hardware [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1053.435951] env[61970]: DEBUG nova.virt.hardware [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1053.436138] env[61970]: DEBUG nova.virt.hardware [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1053.436306] env[61970]: DEBUG nova.virt.hardware [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1053.436482] env[61970]: DEBUG nova.virt.hardware [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1053.437438] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee1cc88c-9248-4585-9c96-33207984bf03 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.445719] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd195fd-79e0-46f9-8910-e410c7d3381a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.613650] env[61970]: DEBUG oslo_concurrency.lockutils [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.240s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.614497] env[61970]: DEBUG nova.compute.manager [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1053.619761] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.121s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.620142] env[61970]: DEBUG nova.objects.instance [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lazy-loading 'resources' on Instance uuid 7f59000a-94d9-45b6-aa7a-300d95793615 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1053.623545] env[61970]: DEBUG nova.compute.manager [req-f0529c52-46dc-4f5c-b68c-12602acb9426 req-e721bc95-8797-4896-bd83-5911638dd127 service nova] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Received event network-vif-deleted-4e65fd7e-c508-4d66-a854-d26fd273014a {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1053.623545] env[61970]: INFO nova.compute.manager [req-f0529c52-46dc-4f5c-b68c-12602acb9426 req-e721bc95-8797-4896-bd83-5911638dd127 service nova] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Neutron deleted interface 4e65fd7e-c508-4d66-a854-d26fd273014a; detaching it from the instance and deleting it from the info cache [ 1053.623545] env[61970]: DEBUG nova.network.neutron [req-f0529c52-46dc-4f5c-b68c-12602acb9426 req-e721bc95-8797-4896-bd83-5911638dd127 service nova] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.885536] env[61970]: DEBUG oslo_concurrency.lockutils [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "31f9d37b-f4ff-4a8b-9477-8253f8305020" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.885789] env[61970]: DEBUG oslo_concurrency.lockutils [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "31f9d37b-f4ff-4a8b-9477-8253f8305020" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.913760] env[61970]: DEBUG oslo_vmware.api [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': task-1356122, 'name': PowerOffVM_Task, 'duration_secs': 0.243937} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.914063] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1053.914243] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1053.914522] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-512a06dc-c1fa-4c78-b2ff-60829622c6c5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.984817] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1053.985737] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1053.985737] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Deleting the datastore file [datastore2] 444939f3-f0d7-4af6-a3fa-40c0a8459839 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1053.985737] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-882d9c88-bb8d-43cb-8975-6b9f4fd722e6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.993358] env[61970]: DEBUG oslo_vmware.api [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Waiting for the task: (returnval){ [ 1053.993358] env[61970]: value = "task-1356124" [ 1053.993358] env[61970]: _type = "Task" [ 1053.993358] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.004180] env[61970]: DEBUG oslo_vmware.api [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': task-1356124, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.071627] env[61970]: DEBUG nova.network.neutron [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Successfully updated port: a7b3f775-a79c-488c-bec7-e8a2b425c91f {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1054.091704] env[61970]: DEBUG nova.network.neutron [-] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.129420] env[61970]: DEBUG nova.compute.utils [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1054.131239] env[61970]: DEBUG nova.compute.manager [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1054.131897] env[61970]: DEBUG nova.network.neutron [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1054.137119] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ec202baf-019b-45fc-92fe-57178a437b62 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.150598] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f16d971-4320-4090-9a56-9f6384213c3f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.181708] env[61970]: DEBUG nova.policy [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '31157f59ec9440989c5f8a7530a752f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f51640cd8b34229b49b947eca804fda', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 1054.197033] env[61970]: DEBUG nova.compute.manager [req-f0529c52-46dc-4f5c-b68c-12602acb9426 req-e721bc95-8797-4896-bd83-5911638dd127 service nova] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Detach interface failed, port_id=4e65fd7e-c508-4d66-a854-d26fd273014a, reason: Instance bc948a41-1a22-46fa-a76a-50b4c4afa4b7 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1054.215354] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "ebb92a5b-f635-4504-a57f-395de514015a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.215610] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "ebb92a5b-f635-4504-a57f-395de514015a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.215820] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "ebb92a5b-f635-4504-a57f-395de514015a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.216014] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "ebb92a5b-f635-4504-a57f-395de514015a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.216201] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "ebb92a5b-f635-4504-a57f-395de514015a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.218356] env[61970]: INFO nova.compute.manager [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Terminating instance [ 1054.220164] env[61970]: DEBUG nova.compute.manager [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1054.220365] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1054.221203] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd2670a4-7e67-4658-944d-d034f2cbff87 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.229412] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1054.229697] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00d201e0-97e5-4db7-92f5-9b017f1f115d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.235969] env[61970]: DEBUG oslo_vmware.api [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1054.235969] env[61970]: value = "task-1356125" [ 1054.235969] env[61970]: _type = "Task" [ 1054.235969] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.244283] env[61970]: DEBUG oslo_vmware.api [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356125, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.305319] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca7edd3e-45d5-458e-aac1-8ae796e1fbeb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.314872] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d783deff-0310-4973-9c09-0f6bf4d745ed {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.345176] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f06628dc-f31e-45d0-ace3-3dc7eb4412d0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.354698] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ef81ea9-43c5-4cb1-b590-18e87907835e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.371684] env[61970]: DEBUG nova.compute.provider_tree [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1054.389263] env[61970]: DEBUG nova.compute.utils [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1054.487562] env[61970]: DEBUG nova.network.neutron [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Successfully created port: 572f62fb-dd55-4cd7-8c2f-a78ef73eb69f {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1054.504438] env[61970]: DEBUG oslo_vmware.api [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Task: {'id': task-1356124, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210881} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.504581] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1054.504808] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1054.505017] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1054.505670] env[61970]: INFO nova.compute.manager [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1054.505670] env[61970]: DEBUG oslo.service.loopingcall [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1054.505670] env[61970]: DEBUG nova.compute.manager [-] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1054.505836] env[61970]: DEBUG nova.network.neutron [-] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1054.575851] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "refresh_cache-d96208cf-fcc5-4014-ac23-1d56ea97050a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.576130] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquired lock "refresh_cache-d96208cf-fcc5-4014-ac23-1d56ea97050a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.576198] env[61970]: DEBUG nova.network.neutron [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1054.594346] env[61970]: INFO nova.compute.manager [-] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Took 1.49 seconds to deallocate network for instance. [ 1054.637568] env[61970]: DEBUG nova.compute.manager [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1054.755991] env[61970]: DEBUG oslo_vmware.api [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356125, 'name': PowerOffVM_Task, 'duration_secs': 0.20211} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.756312] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1054.756546] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1054.756840] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-177408d7-c291-4ea4-9697-94a6a12ea9dd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.820306] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1054.820495] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1054.820751] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Deleting the datastore file [datastore2] ebb92a5b-f635-4504-a57f-395de514015a {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1054.820964] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0fef0ab5-0adf-4c43-993b-9e0e213ffd43 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.827967] env[61970]: DEBUG oslo_vmware.api [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1054.827967] env[61970]: value = "task-1356127" [ 1054.827967] env[61970]: _type = "Task" [ 1054.827967] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.836359] env[61970]: DEBUG oslo_vmware.api [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356127, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.876071] env[61970]: DEBUG nova.scheduler.client.report [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1054.892807] env[61970]: DEBUG oslo_concurrency.lockutils [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "31f9d37b-f4ff-4a8b-9477-8253f8305020" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.102422] env[61970]: DEBUG oslo_concurrency.lockutils [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.109731] env[61970]: DEBUG nova.network.neutron [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1055.239995] env[61970]: DEBUG nova.network.neutron [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Updating instance_info_cache with network_info: [{"id": "a7b3f775-a79c-488c-bec7-e8a2b425c91f", "address": "fa:16:3e:d1:67:be", "network": {"id": "1f53a96b-8a46-41d5-8bf3-759de1fce443", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-108385347-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a44c05a894394d69b8b4fd36c7b2532a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7b3f775-a7", "ovs_interfaceid": "a7b3f775-a79c-488c-bec7-e8a2b425c91f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.338179] env[61970]: DEBUG oslo_vmware.api [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356127, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.253738} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.338384] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1055.338568] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1055.338745] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1055.338922] env[61970]: INFO nova.compute.manager [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1055.339179] env[61970]: DEBUG oslo.service.loopingcall [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1055.339368] env[61970]: DEBUG nova.compute.manager [-] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1055.339482] env[61970]: DEBUG nova.network.neutron [-] [instance: ebb92a5b-f635-4504-a57f-395de514015a] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1055.380890] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.761s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.382959] env[61970]: DEBUG oslo_concurrency.lockutils [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.281s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.383237] env[61970]: DEBUG nova.objects.instance [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lazy-loading 'resources' on Instance uuid bc948a41-1a22-46fa-a76a-50b4c4afa4b7 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.401307] env[61970]: DEBUG nova.network.neutron [-] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.411701] env[61970]: INFO nova.scheduler.client.report [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Deleted allocations for instance 7f59000a-94d9-45b6-aa7a-300d95793615 [ 1055.650713] env[61970]: DEBUG nova.compute.manager [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1055.658396] env[61970]: DEBUG nova.compute.manager [req-521cfc14-8444-4cc4-b83e-4df373a6eb8d req-f25f8874-0674-469d-ba98-c057331eb5e0 service nova] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Received event network-vif-plugged-a7b3f775-a79c-488c-bec7-e8a2b425c91f {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1055.658687] env[61970]: DEBUG oslo_concurrency.lockutils [req-521cfc14-8444-4cc4-b83e-4df373a6eb8d req-f25f8874-0674-469d-ba98-c057331eb5e0 service nova] Acquiring lock "d96208cf-fcc5-4014-ac23-1d56ea97050a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.658945] env[61970]: DEBUG oslo_concurrency.lockutils [req-521cfc14-8444-4cc4-b83e-4df373a6eb8d req-f25f8874-0674-469d-ba98-c057331eb5e0 service nova] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.659185] env[61970]: DEBUG oslo_concurrency.lockutils [req-521cfc14-8444-4cc4-b83e-4df373a6eb8d req-f25f8874-0674-469d-ba98-c057331eb5e0 service nova] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.659412] env[61970]: DEBUG nova.compute.manager [req-521cfc14-8444-4cc4-b83e-4df373a6eb8d req-f25f8874-0674-469d-ba98-c057331eb5e0 service nova] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] No waiting events found dispatching network-vif-plugged-a7b3f775-a79c-488c-bec7-e8a2b425c91f {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1055.659673] env[61970]: WARNING nova.compute.manager [req-521cfc14-8444-4cc4-b83e-4df373a6eb8d req-f25f8874-0674-469d-ba98-c057331eb5e0 service nova] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Received unexpected event network-vif-plugged-a7b3f775-a79c-488c-bec7-e8a2b425c91f for instance with vm_state building and task_state spawning. [ 1055.660442] env[61970]: DEBUG nova.compute.manager [req-521cfc14-8444-4cc4-b83e-4df373a6eb8d req-f25f8874-0674-469d-ba98-c057331eb5e0 service nova] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Received event network-changed-a7b3f775-a79c-488c-bec7-e8a2b425c91f {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1055.660442] env[61970]: DEBUG nova.compute.manager [req-521cfc14-8444-4cc4-b83e-4df373a6eb8d req-f25f8874-0674-469d-ba98-c057331eb5e0 service nova] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Refreshing instance network info cache due to event network-changed-a7b3f775-a79c-488c-bec7-e8a2b425c91f. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1055.660442] env[61970]: DEBUG oslo_concurrency.lockutils [req-521cfc14-8444-4cc4-b83e-4df373a6eb8d req-f25f8874-0674-469d-ba98-c057331eb5e0 service nova] Acquiring lock "refresh_cache-d96208cf-fcc5-4014-ac23-1d56ea97050a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.681993] env[61970]: DEBUG nova.virt.hardware [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1055.682309] env[61970]: DEBUG nova.virt.hardware [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1055.682513] env[61970]: DEBUG nova.virt.hardware [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1055.682737] env[61970]: DEBUG nova.virt.hardware [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1055.682921] env[61970]: DEBUG nova.virt.hardware [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1055.683116] env[61970]: DEBUG nova.virt.hardware [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1055.683367] env[61970]: DEBUG nova.virt.hardware [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1055.683567] env[61970]: DEBUG nova.virt.hardware [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1055.683975] env[61970]: DEBUG nova.virt.hardware [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1055.684245] env[61970]: DEBUG nova.virt.hardware [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1055.684468] env[61970]: DEBUG nova.virt.hardware [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1055.685565] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76600803-7b61-4c9b-8ed4-d2e4a7cadf5a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.694599] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e41df57-51fc-46d8-b096-9cd89b72bc1a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.742343] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Releasing lock "refresh_cache-d96208cf-fcc5-4014-ac23-1d56ea97050a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.742703] env[61970]: DEBUG nova.compute.manager [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Instance network_info: |[{"id": "a7b3f775-a79c-488c-bec7-e8a2b425c91f", "address": "fa:16:3e:d1:67:be", "network": {"id": "1f53a96b-8a46-41d5-8bf3-759de1fce443", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-108385347-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a44c05a894394d69b8b4fd36c7b2532a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7b3f775-a7", "ovs_interfaceid": "a7b3f775-a79c-488c-bec7-e8a2b425c91f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1055.743355] env[61970]: DEBUG oslo_concurrency.lockutils [req-521cfc14-8444-4cc4-b83e-4df373a6eb8d req-f25f8874-0674-469d-ba98-c057331eb5e0 service nova] Acquired lock "refresh_cache-d96208cf-fcc5-4014-ac23-1d56ea97050a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.743544] env[61970]: DEBUG nova.network.neutron [req-521cfc14-8444-4cc4-b83e-4df373a6eb8d req-f25f8874-0674-469d-ba98-c057331eb5e0 service nova] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Refreshing network info cache for port a7b3f775-a79c-488c-bec7-e8a2b425c91f {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1055.744937] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:67:be', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9bb629cd-6d0f-4bed-965c-bd04a2f3ec49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a7b3f775-a79c-488c-bec7-e8a2b425c91f', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1055.753487] env[61970]: DEBUG oslo.service.loopingcall [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1055.758026] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1055.758026] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-502ce7a0-7009-4b55-beb6-1c6a70d10668 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.788089] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1055.788089] env[61970]: value = "task-1356128" [ 1055.788089] env[61970]: _type = "Task" [ 1055.788089] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.801321] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356128, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.906694] env[61970]: INFO nova.compute.manager [-] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Took 1.40 seconds to deallocate network for instance. [ 1055.920782] env[61970]: DEBUG oslo_concurrency.lockutils [None req-dbcc8720-1592-43e8-bb68-39a3a070b698 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "7f59000a-94d9-45b6-aa7a-300d95793615" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.481s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.968978] env[61970]: DEBUG oslo_concurrency.lockutils [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "31f9d37b-f4ff-4a8b-9477-8253f8305020" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.969255] env[61970]: DEBUG oslo_concurrency.lockutils [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "31f9d37b-f4ff-4a8b-9477-8253f8305020" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.969518] env[61970]: INFO nova.compute.manager [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Attaching volume f1fea832-29d5-441d-a553-3c04a1903995 to /dev/sdb [ 1056.016417] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1f0c90-d687-47c4-9a19-71ec1161509c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.022415] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced5236e-f4a7-4b60-84ab-2cd9a354d5e5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.029722] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e786781b-6376-4565-a79d-814a359f9165 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.039992] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e936f2f0-ef0e-479d-b7b4-32966ee9fbf9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.066556] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1da348d-2e84-4b54-9a93-6c339072b436 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.074605] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28e4b41d-4800-4e51-8e28-63ae3c619cc6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.081237] env[61970]: DEBUG nova.virt.block_device [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Updating existing volume attachment record: 28ea9ea8-d076-48b3-8875-9d51215b5cfd {{(pid=61970) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1056.093475] env[61970]: DEBUG nova.compute.provider_tree [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.112544] env[61970]: DEBUG nova.network.neutron [req-521cfc14-8444-4cc4-b83e-4df373a6eb8d req-f25f8874-0674-469d-ba98-c057331eb5e0 service nova] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Updated VIF entry in instance network info cache for port a7b3f775-a79c-488c-bec7-e8a2b425c91f. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1056.112900] env[61970]: DEBUG nova.network.neutron [req-521cfc14-8444-4cc4-b83e-4df373a6eb8d req-f25f8874-0674-469d-ba98-c057331eb5e0 service nova] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Updating instance_info_cache with network_info: [{"id": "a7b3f775-a79c-488c-bec7-e8a2b425c91f", "address": "fa:16:3e:d1:67:be", "network": {"id": "1f53a96b-8a46-41d5-8bf3-759de1fce443", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-108385347-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a44c05a894394d69b8b4fd36c7b2532a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7b3f775-a7", "ovs_interfaceid": "a7b3f775-a79c-488c-bec7-e8a2b425c91f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.271468] env[61970]: DEBUG nova.network.neutron [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Successfully updated port: 572f62fb-dd55-4cd7-8c2f-a78ef73eb69f {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1056.285041] env[61970]: DEBUG nova.network.neutron [-] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.304616] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356128, 'name': CreateVM_Task, 'duration_secs': 0.323079} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.304980] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1056.305711] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.305894] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.307614] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1056.307614] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2896026-cc00-4402-9afe-2f563e8b1164 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.314858] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1056.314858] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b24cec-630d-6543-5dff-7212ae8288fa" [ 1056.314858] env[61970]: _type = "Task" [ 1056.314858] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.325754] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b24cec-630d-6543-5dff-7212ae8288fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.414217] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.596182] env[61970]: DEBUG nova.scheduler.client.report [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1056.615151] env[61970]: DEBUG oslo_concurrency.lockutils [req-521cfc14-8444-4cc4-b83e-4df373a6eb8d req-f25f8874-0674-469d-ba98-c057331eb5e0 service nova] Releasing lock "refresh_cache-d96208cf-fcc5-4014-ac23-1d56ea97050a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.615406] env[61970]: DEBUG nova.compute.manager [req-521cfc14-8444-4cc4-b83e-4df373a6eb8d req-f25f8874-0674-469d-ba98-c057331eb5e0 service nova] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Received event network-vif-deleted-b8177d93-2145-46f4-b212-1b5bd5c97e41 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1056.773335] env[61970]: DEBUG oslo_concurrency.lockutils [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Acquiring lock "refresh_cache-ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.773680] env[61970]: DEBUG oslo_concurrency.lockutils [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Acquired lock "refresh_cache-ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.773723] env[61970]: DEBUG nova.network.neutron [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1056.785385] env[61970]: INFO nova.compute.manager [-] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Took 1.45 seconds to deallocate network for instance. [ 1056.825757] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b24cec-630d-6543-5dff-7212ae8288fa, 'name': SearchDatastore_Task, 'duration_secs': 0.044443} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.826092] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.826337] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1056.826601] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.826728] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.826906] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1056.827204] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4bb73c3b-bd78-4ed1-b426-7a573690711d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.839838] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1056.840033] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1056.840741] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9faae6b8-91e0-4058-a094-57174e62ff53 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.846125] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1056.846125] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e2dd9d-7f38-261c-c547-ac92b6245968" [ 1056.846125] env[61970]: _type = "Task" [ 1056.846125] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.853901] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e2dd9d-7f38-261c-c547-ac92b6245968, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.102250] env[61970]: DEBUG oslo_concurrency.lockutils [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.719s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.104581] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.691s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.104825] env[61970]: DEBUG nova.objects.instance [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Lazy-loading 'resources' on Instance uuid 444939f3-f0d7-4af6-a3fa-40c0a8459839 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1057.128657] env[61970]: INFO nova.scheduler.client.report [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleted allocations for instance bc948a41-1a22-46fa-a76a-50b4c4afa4b7 [ 1057.295205] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.304431] env[61970]: DEBUG nova.network.neutron [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1057.357501] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e2dd9d-7f38-261c-c547-ac92b6245968, 'name': SearchDatastore_Task, 'duration_secs': 0.017241} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.358352] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aefe0bf4-2d6e-4bcc-a010-bd4a7f50eb3b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.363309] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1057.363309] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]521bfbbb-3a4b-fa6d-7ca4-d1db9ca84d43" [ 1057.363309] env[61970]: _type = "Task" [ 1057.363309] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.372879] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]521bfbbb-3a4b-fa6d-7ca4-d1db9ca84d43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.469853] env[61970]: DEBUG nova.network.neutron [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Updating instance_info_cache with network_info: [{"id": "572f62fb-dd55-4cd7-8c2f-a78ef73eb69f", "address": "fa:16:3e:b8:b0:65", "network": {"id": "944b5b23-1fb7-4395-830f-388fdb4b010a", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1849890351-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f51640cd8b34229b49b947eca804fda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap572f62fb-dd", "ovs_interfaceid": "572f62fb-dd55-4cd7-8c2f-a78ef73eb69f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.637793] env[61970]: DEBUG oslo_concurrency.lockutils [None req-70b84822-49ae-433a-9eb5-24dbfbc3e5dd tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "bc948a41-1a22-46fa-a76a-50b4c4afa4b7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.675s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.685095] env[61970]: DEBUG nova.compute.manager [req-2cbeb49c-08bb-4d87-8d52-b26112308cf7 req-de9c305a-0a64-499b-8e63-6bc26a647de0 service nova] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Received event network-vif-deleted-9a1c8b46-0249-4e0b-94de-b0d0062ca69c {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1057.685313] env[61970]: DEBUG nova.compute.manager [req-2cbeb49c-08bb-4d87-8d52-b26112308cf7 req-de9c305a-0a64-499b-8e63-6bc26a647de0 service nova] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Received event network-vif-plugged-572f62fb-dd55-4cd7-8c2f-a78ef73eb69f {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1057.685515] env[61970]: DEBUG oslo_concurrency.lockutils [req-2cbeb49c-08bb-4d87-8d52-b26112308cf7 req-de9c305a-0a64-499b-8e63-6bc26a647de0 service nova] Acquiring lock "ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.685719] env[61970]: DEBUG oslo_concurrency.lockutils [req-2cbeb49c-08bb-4d87-8d52-b26112308cf7 req-de9c305a-0a64-499b-8e63-6bc26a647de0 service nova] Lock "ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.685946] env[61970]: DEBUG oslo_concurrency.lockutils [req-2cbeb49c-08bb-4d87-8d52-b26112308cf7 req-de9c305a-0a64-499b-8e63-6bc26a647de0 service nova] Lock "ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.686178] env[61970]: DEBUG nova.compute.manager [req-2cbeb49c-08bb-4d87-8d52-b26112308cf7 req-de9c305a-0a64-499b-8e63-6bc26a647de0 service nova] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] No waiting events found dispatching network-vif-plugged-572f62fb-dd55-4cd7-8c2f-a78ef73eb69f {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1057.686369] env[61970]: WARNING nova.compute.manager [req-2cbeb49c-08bb-4d87-8d52-b26112308cf7 req-de9c305a-0a64-499b-8e63-6bc26a647de0 service nova] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Received unexpected event network-vif-plugged-572f62fb-dd55-4cd7-8c2f-a78ef73eb69f for instance with vm_state building and task_state spawning. [ 1057.686569] env[61970]: DEBUG nova.compute.manager [req-2cbeb49c-08bb-4d87-8d52-b26112308cf7 req-de9c305a-0a64-499b-8e63-6bc26a647de0 service nova] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Received event network-changed-572f62fb-dd55-4cd7-8c2f-a78ef73eb69f {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1057.686799] env[61970]: DEBUG nova.compute.manager [req-2cbeb49c-08bb-4d87-8d52-b26112308cf7 req-de9c305a-0a64-499b-8e63-6bc26a647de0 service nova] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Refreshing instance network info cache due to event network-changed-572f62fb-dd55-4cd7-8c2f-a78ef73eb69f. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1057.686994] env[61970]: DEBUG oslo_concurrency.lockutils [req-2cbeb49c-08bb-4d87-8d52-b26112308cf7 req-de9c305a-0a64-499b-8e63-6bc26a647de0 service nova] Acquiring lock "refresh_cache-ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.698646] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee437a0b-cd5a-4b1c-bc41-b8c0b1c31e7c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.706069] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fffdf07-1151-4151-8d4e-08fb91eeb9a4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.735916] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df5e939d-a84b-4ea3-9fc0-ef427cacdcfe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.742902] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d97afc35-d712-4572-bb2d-8689f0527326 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.755761] env[61970]: DEBUG nova.compute.provider_tree [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1057.826084] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.826338] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.874013] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]521bfbbb-3a4b-fa6d-7ca4-d1db9ca84d43, 'name': SearchDatastore_Task, 'duration_secs': 0.010314} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.874269] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.874528] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] d96208cf-fcc5-4014-ac23-1d56ea97050a/d96208cf-fcc5-4014-ac23-1d56ea97050a.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1057.874780] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-43566e4f-1aca-40a6-ae05-f5e3cfa10e3b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.881671] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1057.881671] env[61970]: value = "task-1356130" [ 1057.881671] env[61970]: _type = "Task" [ 1057.881671] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.888960] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356130, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.973159] env[61970]: DEBUG oslo_concurrency.lockutils [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Releasing lock "refresh_cache-ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.973461] env[61970]: DEBUG nova.compute.manager [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Instance network_info: |[{"id": "572f62fb-dd55-4cd7-8c2f-a78ef73eb69f", "address": "fa:16:3e:b8:b0:65", "network": {"id": "944b5b23-1fb7-4395-830f-388fdb4b010a", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1849890351-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f51640cd8b34229b49b947eca804fda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap572f62fb-dd", "ovs_interfaceid": "572f62fb-dd55-4cd7-8c2f-a78ef73eb69f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1057.973785] env[61970]: DEBUG oslo_concurrency.lockutils [req-2cbeb49c-08bb-4d87-8d52-b26112308cf7 req-de9c305a-0a64-499b-8e63-6bc26a647de0 service nova] Acquired lock "refresh_cache-ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.973985] env[61970]: DEBUG nova.network.neutron [req-2cbeb49c-08bb-4d87-8d52-b26112308cf7 req-de9c305a-0a64-499b-8e63-6bc26a647de0 service nova] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Refreshing network info cache for port 572f62fb-dd55-4cd7-8c2f-a78ef73eb69f {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1057.975304] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:b0:65', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0f925dc8-2145-457e-a4d4-c07117356dd0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '572f62fb-dd55-4cd7-8c2f-a78ef73eb69f', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1057.982909] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Creating folder: Project (4f51640cd8b34229b49b947eca804fda). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1057.983873] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ec04c124-e4a5-4740-9566-bc0445e0eda2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.993450] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Created folder: Project (4f51640cd8b34229b49b947eca804fda) in parent group-v288740. [ 1057.993658] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Creating folder: Instances. Parent ref: group-v288875. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1057.993901] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-afa8b3e6-d30f-49dd-a724-f52666bfa6ee {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.001878] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Created folder: Instances in parent group-v288875. [ 1058.002126] env[61970]: DEBUG oslo.service.loopingcall [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1058.002328] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1058.002540] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2cb096c7-f848-4c5d-be78-c3fd22db1729 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.022389] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1058.022389] env[61970]: value = "task-1356133" [ 1058.022389] env[61970]: _type = "Task" [ 1058.022389] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.031547] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356133, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.259094] env[61970]: DEBUG nova.scheduler.client.report [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1058.328602] env[61970]: DEBUG nova.compute.manager [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1058.392254] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356130, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.452779} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.392775] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] d96208cf-fcc5-4014-ac23-1d56ea97050a/d96208cf-fcc5-4014-ac23-1d56ea97050a.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1058.392880] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1058.393094] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ce646bb8-953a-4685-a367-62531610624b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.399350] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1058.399350] env[61970]: value = "task-1356135" [ 1058.399350] env[61970]: _type = "Task" [ 1058.399350] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.408207] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356135, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.533798] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356133, 'name': CreateVM_Task, 'duration_secs': 0.348407} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.533975] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1058.534660] env[61970]: DEBUG oslo_concurrency.lockutils [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.534886] env[61970]: DEBUG oslo_concurrency.lockutils [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.535181] env[61970]: DEBUG oslo_concurrency.lockutils [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1058.535446] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-691babb4-572b-4577-95c0-9af4cc98292c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.541631] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Waiting for the task: (returnval){ [ 1058.541631] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ae9641-a53b-5d97-d2e3-e3be6e7ebbd7" [ 1058.541631] env[61970]: _type = "Task" [ 1058.541631] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.550440] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ae9641-a53b-5d97-d2e3-e3be6e7ebbd7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.720653] env[61970]: DEBUG nova.network.neutron [req-2cbeb49c-08bb-4d87-8d52-b26112308cf7 req-de9c305a-0a64-499b-8e63-6bc26a647de0 service nova] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Updated VIF entry in instance network info cache for port 572f62fb-dd55-4cd7-8c2f-a78ef73eb69f. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1058.721375] env[61970]: DEBUG nova.network.neutron [req-2cbeb49c-08bb-4d87-8d52-b26112308cf7 req-de9c305a-0a64-499b-8e63-6bc26a647de0 service nova] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Updating instance_info_cache with network_info: [{"id": "572f62fb-dd55-4cd7-8c2f-a78ef73eb69f", "address": "fa:16:3e:b8:b0:65", "network": {"id": "944b5b23-1fb7-4395-830f-388fdb4b010a", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1849890351-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f51640cd8b34229b49b947eca804fda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap572f62fb-dd", "ovs_interfaceid": "572f62fb-dd55-4cd7-8c2f-a78ef73eb69f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.764999] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.660s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.767239] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.472s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.767455] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.791654] env[61970]: INFO nova.scheduler.client.report [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Deleted allocations for instance 444939f3-f0d7-4af6-a3fa-40c0a8459839 [ 1058.793343] env[61970]: INFO nova.scheduler.client.report [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Deleted allocations for instance ebb92a5b-f635-4504-a57f-395de514015a [ 1058.851283] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.851594] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.853148] env[61970]: INFO nova.compute.claims [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1058.909136] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356135, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057299} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.909496] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1058.910248] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422a7940-530e-4f2c-88f8-35c809be177a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.932824] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] d96208cf-fcc5-4014-ac23-1d56ea97050a/d96208cf-fcc5-4014-ac23-1d56ea97050a.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1058.933135] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0dd453f-1c39-446e-be66-bb9db9077ace {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.955427] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1058.955427] env[61970]: value = "task-1356136" [ 1058.955427] env[61970]: _type = "Task" [ 1058.955427] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.964053] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356136, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.052352] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ae9641-a53b-5d97-d2e3-e3be6e7ebbd7, 'name': SearchDatastore_Task, 'duration_secs': 0.014805} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.052782] env[61970]: DEBUG oslo_concurrency.lockutils [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.053057] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1059.053322] env[61970]: DEBUG oslo_concurrency.lockutils [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.053491] env[61970]: DEBUG oslo_concurrency.lockutils [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.053769] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1059.053972] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3ca9bd16-6b06-4802-a854-d74dfca0dab0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.061524] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1059.061747] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1059.062522] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11795206-29b8-4dda-9142-02a728794a34 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.067249] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Waiting for the task: (returnval){ [ 1059.067249] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52052239-00c4-eda9-e63a-82c21c25dd3d" [ 1059.067249] env[61970]: _type = "Task" [ 1059.067249] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.074767] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52052239-00c4-eda9-e63a-82c21c25dd3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.224343] env[61970]: DEBUG oslo_concurrency.lockutils [req-2cbeb49c-08bb-4d87-8d52-b26112308cf7 req-de9c305a-0a64-499b-8e63-6bc26a647de0 service nova] Releasing lock "refresh_cache-ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.304142] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bba56994-28cd-4089-9031-799a6e2a308a tempest-ServersNegativeTestMultiTenantJSON-1918592166 tempest-ServersNegativeTestMultiTenantJSON-1918592166-project-member] Lock "444939f3-f0d7-4af6-a3fa-40c0a8459839" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.923s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.305155] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1c1fb4e2-eb1a-4cc9-b1d9-238babf09371 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "ebb92a5b-f635-4504-a57f-395de514015a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.089s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.465258] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356136, 'name': ReconfigVM_Task, 'duration_secs': 0.243222} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.465545] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Reconfigured VM instance instance-00000064 to attach disk [datastore1] d96208cf-fcc5-4014-ac23-1d56ea97050a/d96208cf-fcc5-4014-ac23-1d56ea97050a.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1059.466209] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d035f026-0578-40e5-b781-5a1971080059 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.472316] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1059.472316] env[61970]: value = "task-1356137" [ 1059.472316] env[61970]: _type = "Task" [ 1059.472316] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.482634] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356137, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.582035] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52052239-00c4-eda9-e63a-82c21c25dd3d, 'name': SearchDatastore_Task, 'duration_secs': 0.007866} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.582035] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be28d660-c187-449a-8291-e7fc71ee1857 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.584707] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Waiting for the task: (returnval){ [ 1059.584707] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52040733-939d-3a6c-8825-a30792f290e6" [ 1059.584707] env[61970]: _type = "Task" [ 1059.584707] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.593819] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52040733-939d-3a6c-8825-a30792f290e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.789357] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "183133c6-7e4e-4dd4-b8e0-cb4f8338264c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.789648] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "183133c6-7e4e-4dd4-b8e0-cb4f8338264c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.950191] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937acd8b-3f92-438b-a7be-c2ed9cd7d6b4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.958062] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7422526-1501-4893-b849-a76faa70000f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.991075] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8135d85-ae40-4f55-8716-589c4baf8777 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.000035] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356137, 'name': Rename_Task, 'duration_secs': 0.136117} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.001456] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1060.001456] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c206cdf-9a41-4cf1-aa81-6b4d8d65b594 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.003192] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fda8250-2704-47f8-a751-6eeb5e667ae7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.016296] env[61970]: DEBUG nova.compute.provider_tree [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1060.018575] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1060.018575] env[61970]: value = "task-1356138" [ 1060.018575] env[61970]: _type = "Task" [ 1060.018575] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.026439] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356138, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.097258] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52040733-939d-3a6c-8825-a30792f290e6, 'name': SearchDatastore_Task, 'duration_secs': 0.036468} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.097531] env[61970]: DEBUG oslo_concurrency.lockutils [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.097799] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97/ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1060.098084] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4e1e2db5-1eb5-4456-8f2f-ce961753914f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.103858] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Waiting for the task: (returnval){ [ 1060.103858] env[61970]: value = "task-1356139" [ 1060.103858] env[61970]: _type = "Task" [ 1060.103858] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.110933] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': task-1356139, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.291998] env[61970]: DEBUG nova.compute.manager [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1060.520898] env[61970]: DEBUG nova.scheduler.client.report [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1060.539102] env[61970]: DEBUG oslo_vmware.api [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356138, 'name': PowerOnVM_Task, 'duration_secs': 0.438289} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.540508] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1060.541171] env[61970]: INFO nova.compute.manager [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Took 7.14 seconds to spawn the instance on the hypervisor. [ 1060.545022] env[61970]: DEBUG nova.compute.manager [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1060.545022] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7aac265-73e2-49ff-a94a-4618a777e948 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.553783] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "5a8cdb36-1845-4806-a6bc-22afb47886e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.554012] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "5a8cdb36-1845-4806-a6bc-22afb47886e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.616237] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': task-1356139, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.640977] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Volume attach. Driver type: vmdk {{(pid=61970) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1060.641265] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288874', 'volume_id': 'f1fea832-29d5-441d-a553-3c04a1903995', 'name': 'volume-f1fea832-29d5-441d-a553-3c04a1903995', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '31f9d37b-f4ff-4a8b-9477-8253f8305020', 'attached_at': '', 'detached_at': '', 'volume_id': 'f1fea832-29d5-441d-a553-3c04a1903995', 'serial': 'f1fea832-29d5-441d-a553-3c04a1903995'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1060.642210] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac59406-177a-406b-a7af-7b5501932949 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.659971] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e377101-d326-4ee6-98d5-0448a7733d34 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.685455] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] volume-f1fea832-29d5-441d-a553-3c04a1903995/volume-f1fea832-29d5-441d-a553-3c04a1903995.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1060.686180] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f5e63a7-85de-4282-9f1e-2be2fdc4c391 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.705813] env[61970]: DEBUG oslo_vmware.api [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1060.705813] env[61970]: value = "task-1356140" [ 1060.705813] env[61970]: _type = "Task" [ 1060.705813] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.713793] env[61970]: DEBUG oslo_vmware.api [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356140, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.816976] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.030884] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.179s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.031544] env[61970]: DEBUG nova.compute.manager [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1061.034390] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.218s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.035789] env[61970]: INFO nova.compute.claims [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1061.061274] env[61970]: DEBUG nova.compute.manager [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1061.070375] env[61970]: INFO nova.compute.manager [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Took 14.46 seconds to build instance. [ 1061.118028] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': task-1356139, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.558945} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.118630] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97/ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1061.118921] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1061.119251] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a3b71d11-771e-4210-9636-5ae1c39c4a8b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.126345] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Waiting for the task: (returnval){ [ 1061.126345] env[61970]: value = "task-1356141" [ 1061.126345] env[61970]: _type = "Task" [ 1061.126345] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.136127] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': task-1356141, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.218360] env[61970]: DEBUG oslo_vmware.api [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356140, 'name': ReconfigVM_Task, 'duration_secs': 0.351953} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.218360] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Reconfigured VM instance instance-00000061 to attach disk [datastore1] volume-f1fea832-29d5-441d-a553-3c04a1903995/volume-f1fea832-29d5-441d-a553-3c04a1903995.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1061.220481] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-200bdcaf-6702-4f84-a890-5534dab236ac {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.235593] env[61970]: DEBUG oslo_vmware.api [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1061.235593] env[61970]: value = "task-1356142" [ 1061.235593] env[61970]: _type = "Task" [ 1061.235593] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.243487] env[61970]: DEBUG oslo_vmware.api [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356142, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.542925] env[61970]: DEBUG nova.compute.utils [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1061.546105] env[61970]: DEBUG nova.compute.manager [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1061.546291] env[61970]: DEBUG nova.network.neutron [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1061.573517] env[61970]: DEBUG oslo_concurrency.lockutils [None req-1bbf2997-1d9e-498e-8989-31771c405384 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.976s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.593326] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.600228] env[61970]: DEBUG nova.policy [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf68cfc72ee04882ad02b827751520bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1d4724cbb0841859369af01f2c03db1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 1061.637943] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': task-1356141, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0606} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.638229] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1061.638987] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-952f13a2-ddae-42b8-84f9-910b2e187936 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.661407] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97/ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1061.661682] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-820e1e70-250c-4e53-962b-fc9b09294577 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.684812] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Waiting for the task: (returnval){ [ 1061.684812] env[61970]: value = "task-1356143" [ 1061.684812] env[61970]: _type = "Task" [ 1061.684812] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.692973] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': task-1356143, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.746079] env[61970]: DEBUG oslo_vmware.api [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356142, 'name': ReconfigVM_Task, 'duration_secs': 0.137883} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.746395] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288874', 'volume_id': 'f1fea832-29d5-441d-a553-3c04a1903995', 'name': 'volume-f1fea832-29d5-441d-a553-3c04a1903995', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '31f9d37b-f4ff-4a8b-9477-8253f8305020', 'attached_at': '', 'detached_at': '', 'volume_id': 'f1fea832-29d5-441d-a553-3c04a1903995', 'serial': 'f1fea832-29d5-441d-a553-3c04a1903995'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1062.052321] env[61970]: DEBUG nova.compute.manager [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1062.165606] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d659bf-0b99-4f0e-a8b9-6b07ebdb5d84 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.175595] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59fcbe99-a5ff-4e0c-a214-f8f592b3085e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.222591] env[61970]: DEBUG nova.network.neutron [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Successfully created port: 2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1062.225904] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03e96f7c-0791-4635-982f-eb3a9dbdcc04 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.233523] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': task-1356143, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.238430] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3411b813-ee70-4141-88ea-ac70c7ff4db8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.255318] env[61970]: DEBUG nova.compute.provider_tree [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1062.419900] env[61970]: DEBUG nova.compute.manager [req-85d31a05-f89e-40a1-ba4e-1ee03804d426 req-b205e629-d472-4195-b853-efe681d39b75 service nova] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Received event network-changed-a7b3f775-a79c-488c-bec7-e8a2b425c91f {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1062.420158] env[61970]: DEBUG nova.compute.manager [req-85d31a05-f89e-40a1-ba4e-1ee03804d426 req-b205e629-d472-4195-b853-efe681d39b75 service nova] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Refreshing instance network info cache due to event network-changed-a7b3f775-a79c-488c-bec7-e8a2b425c91f. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1062.420408] env[61970]: DEBUG oslo_concurrency.lockutils [req-85d31a05-f89e-40a1-ba4e-1ee03804d426 req-b205e629-d472-4195-b853-efe681d39b75 service nova] Acquiring lock "refresh_cache-d96208cf-fcc5-4014-ac23-1d56ea97050a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.420569] env[61970]: DEBUG oslo_concurrency.lockutils [req-85d31a05-f89e-40a1-ba4e-1ee03804d426 req-b205e629-d472-4195-b853-efe681d39b75 service nova] Acquired lock "refresh_cache-d96208cf-fcc5-4014-ac23-1d56ea97050a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.420742] env[61970]: DEBUG nova.network.neutron [req-85d31a05-f89e-40a1-ba4e-1ee03804d426 req-b205e629-d472-4195-b853-efe681d39b75 service nova] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Refreshing network info cache for port a7b3f775-a79c-488c-bec7-e8a2b425c91f {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1062.698580] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': task-1356143, 'name': ReconfigVM_Task, 'duration_secs': 0.734934} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.698948] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Reconfigured VM instance instance-00000065 to attach disk [datastore1] ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97/ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1062.699644] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93258fda-052d-4bad-82b1-4c0c1aeec7bc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.705878] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Waiting for the task: (returnval){ [ 1062.705878] env[61970]: value = "task-1356144" [ 1062.705878] env[61970]: _type = "Task" [ 1062.705878] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.713302] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': task-1356144, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.759415] env[61970]: DEBUG nova.scheduler.client.report [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1062.789787] env[61970]: DEBUG nova.objects.instance [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lazy-loading 'flavor' on Instance uuid 31f9d37b-f4ff-4a8b-9477-8253f8305020 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1063.063667] env[61970]: DEBUG nova.compute.manager [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1063.094605] env[61970]: DEBUG nova.virt.hardware [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1063.094899] env[61970]: DEBUG nova.virt.hardware [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1063.095202] env[61970]: DEBUG nova.virt.hardware [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1063.095437] env[61970]: DEBUG nova.virt.hardware [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1063.095934] env[61970]: DEBUG nova.virt.hardware [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1063.095934] env[61970]: DEBUG nova.virt.hardware [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1063.096097] env[61970]: DEBUG nova.virt.hardware [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1063.096172] env[61970]: DEBUG nova.virt.hardware [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1063.096343] env[61970]: DEBUG nova.virt.hardware [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1063.096515] env[61970]: DEBUG nova.virt.hardware [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1063.096690] env[61970]: DEBUG nova.virt.hardware [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1063.097635] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228f6e0b-7a0c-40d9-b103-fb809aacb33a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.110293] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14346d99-cc14-44a2-a96e-09dd065d7817 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.216863] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': task-1356144, 'name': Rename_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.243454] env[61970]: DEBUG nova.network.neutron [req-85d31a05-f89e-40a1-ba4e-1ee03804d426 req-b205e629-d472-4195-b853-efe681d39b75 service nova] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Updated VIF entry in instance network info cache for port a7b3f775-a79c-488c-bec7-e8a2b425c91f. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1063.243830] env[61970]: DEBUG nova.network.neutron [req-85d31a05-f89e-40a1-ba4e-1ee03804d426 req-b205e629-d472-4195-b853-efe681d39b75 service nova] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Updating instance_info_cache with network_info: [{"id": "a7b3f775-a79c-488c-bec7-e8a2b425c91f", "address": "fa:16:3e:d1:67:be", "network": {"id": "1f53a96b-8a46-41d5-8bf3-759de1fce443", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-108385347-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a44c05a894394d69b8b4fd36c7b2532a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7b3f775-a7", "ovs_interfaceid": "a7b3f775-a79c-488c-bec7-e8a2b425c91f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.264298] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.230s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.264838] env[61970]: DEBUG nova.compute.manager [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1063.267409] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.674s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.268861] env[61970]: INFO nova.compute.claims [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1063.296539] env[61970]: DEBUG oslo_concurrency.lockutils [None req-cacab6c4-86d9-4078-880b-3f82ac32daa4 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "31f9d37b-f4ff-4a8b-9477-8253f8305020" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.327s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.494089] env[61970]: DEBUG oslo_concurrency.lockutils [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "31f9d37b-f4ff-4a8b-9477-8253f8305020" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.495138] env[61970]: DEBUG oslo_concurrency.lockutils [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "31f9d37b-f4ff-4a8b-9477-8253f8305020" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.717670] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': task-1356144, 'name': Rename_Task, 'duration_secs': 0.927549} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.717965] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1063.718236] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f264e861-4f5e-4307-824b-e85144b17ed1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.724939] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Waiting for the task: (returnval){ [ 1063.724939] env[61970]: value = "task-1356145" [ 1063.724939] env[61970]: _type = "Task" [ 1063.724939] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.731944] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': task-1356145, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.746639] env[61970]: DEBUG oslo_concurrency.lockutils [req-85d31a05-f89e-40a1-ba4e-1ee03804d426 req-b205e629-d472-4195-b853-efe681d39b75 service nova] Releasing lock "refresh_cache-d96208cf-fcc5-4014-ac23-1d56ea97050a" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.774311] env[61970]: DEBUG nova.compute.utils [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1063.775700] env[61970]: DEBUG nova.compute.manager [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1063.775840] env[61970]: DEBUG nova.network.neutron [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1063.839873] env[61970]: DEBUG nova.compute.manager [req-e4dc2566-37d0-4f00-958f-6cd045e259af req-bb007878-0a98-4ace-979a-47dc211123b5 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Received event network-vif-plugged-2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1063.840099] env[61970]: DEBUG oslo_concurrency.lockutils [req-e4dc2566-37d0-4f00-958f-6cd045e259af req-bb007878-0a98-4ace-979a-47dc211123b5 service nova] Acquiring lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.840323] env[61970]: DEBUG oslo_concurrency.lockutils [req-e4dc2566-37d0-4f00-958f-6cd045e259af req-bb007878-0a98-4ace-979a-47dc211123b5 service nova] Lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.840498] env[61970]: DEBUG oslo_concurrency.lockutils [req-e4dc2566-37d0-4f00-958f-6cd045e259af req-bb007878-0a98-4ace-979a-47dc211123b5 service nova] Lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.840668] env[61970]: DEBUG nova.compute.manager [req-e4dc2566-37d0-4f00-958f-6cd045e259af req-bb007878-0a98-4ace-979a-47dc211123b5 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] No waiting events found dispatching network-vif-plugged-2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1063.840835] env[61970]: WARNING nova.compute.manager [req-e4dc2566-37d0-4f00-958f-6cd045e259af req-bb007878-0a98-4ace-979a-47dc211123b5 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Received unexpected event network-vif-plugged-2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 for instance with vm_state building and task_state spawning. [ 1063.865355] env[61970]: DEBUG nova.policy [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '223ba28a2a204f8ba9e6d2e0f876ddd2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '28a078b8af9c469eb279be4da7459166', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 1063.931139] env[61970]: DEBUG nova.network.neutron [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Successfully updated port: 2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1063.997748] env[61970]: INFO nova.compute.manager [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Detaching volume f1fea832-29d5-441d-a553-3c04a1903995 [ 1064.033667] env[61970]: INFO nova.virt.block_device [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Attempting to driver detach volume f1fea832-29d5-441d-a553-3c04a1903995 from mountpoint /dev/sdb [ 1064.033973] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Volume detach. Driver type: vmdk {{(pid=61970) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1064.034332] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288874', 'volume_id': 'f1fea832-29d5-441d-a553-3c04a1903995', 'name': 'volume-f1fea832-29d5-441d-a553-3c04a1903995', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '31f9d37b-f4ff-4a8b-9477-8253f8305020', 'attached_at': '', 'detached_at': '', 'volume_id': 'f1fea832-29d5-441d-a553-3c04a1903995', 'serial': 'f1fea832-29d5-441d-a553-3c04a1903995'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1064.035245] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-098d3b04-0c97-4a9d-915e-f1e3910331f7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.057868] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-743ea804-ccf2-4550-8f9b-09e4bf7a704a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.065233] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de82e0a1-e98b-4944-b8f9-1c6cb365b793 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.088132] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba750c39-3d23-4689-b1f0-d7cbb81e3c18 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.103395] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] The volume has not been displaced from its original location: [datastore1] volume-f1fea832-29d5-441d-a553-3c04a1903995/volume-f1fea832-29d5-441d-a553-3c04a1903995.vmdk. No consolidation needed. {{(pid=61970) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1064.108657] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Reconfiguring VM instance instance-00000061 to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1064.108818] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-767c1be9-f532-41a9-8f34-e725193d86c5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.129734] env[61970]: DEBUG oslo_vmware.api [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1064.129734] env[61970]: value = "task-1356146" [ 1064.129734] env[61970]: _type = "Task" [ 1064.129734] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.139615] env[61970]: DEBUG oslo_vmware.api [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356146, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.151499] env[61970]: DEBUG nova.network.neutron [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Successfully created port: f89f6ca5-913c-4eff-ab7e-2199ee90ccd7 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1064.236419] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': task-1356145, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.281408] env[61970]: DEBUG nova.compute.manager [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1064.392682] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f59a8151-a066-49f9-8042-310d4320e4a9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.399601] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d320f2-5339-4dea-a97a-ae55b073f49d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.432586] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea279405-1312-43d2-b8f8-767b5b42429a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.437501] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.437648] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquired lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.437796] env[61970]: DEBUG nova.network.neutron [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1064.444515] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a40c6e-dc11-4abf-81b5-d4a5b1471505 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.464977] env[61970]: DEBUG nova.compute.provider_tree [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.640068] env[61970]: DEBUG oslo_vmware.api [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356146, 'name': ReconfigVM_Task, 'duration_secs': 0.251071} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.640301] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Reconfigured VM instance instance-00000061 to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1064.645053] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3a6dfe9-edb3-4116-a157-9b3cea861562 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.659637] env[61970]: DEBUG oslo_vmware.api [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1064.659637] env[61970]: value = "task-1356147" [ 1064.659637] env[61970]: _type = "Task" [ 1064.659637] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.667740] env[61970]: DEBUG oslo_vmware.api [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356147, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.736042] env[61970]: DEBUG oslo_vmware.api [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': task-1356145, 'name': PowerOnVM_Task, 'duration_secs': 0.808535} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.736042] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1064.736042] env[61970]: INFO nova.compute.manager [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Took 9.08 seconds to spawn the instance on the hypervisor. [ 1064.736042] env[61970]: DEBUG nova.compute.manager [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1064.736837] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4af21f0-f9c9-41d0-b2e5-6082b27a70cd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.968073] env[61970]: DEBUG nova.scheduler.client.report [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1064.973090] env[61970]: DEBUG nova.network.neutron [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1065.131813] env[61970]: DEBUG nova.network.neutron [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Updating instance_info_cache with network_info: [{"id": "2fa02dfe-2cb0-4a72-bc30-4f991339dcb9", "address": "fa:16:3e:df:19:94", "network": {"id": "70081995-3d6c-466e-b811-1c8e532e552f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-641102750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1d4724cbb0841859369af01f2c03db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aaf1b231-3660-4453-b4f3-44d825b9a5dd", "external-id": "nsx-vlan-transportzone-6", "segmentation_id": 6, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fa02dfe-2c", "ovs_interfaceid": "2fa02dfe-2cb0-4a72-bc30-4f991339dcb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.169286] env[61970]: DEBUG oslo_vmware.api [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356147, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.252429] env[61970]: INFO nova.compute.manager [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Took 15.18 seconds to build instance. [ 1065.293061] env[61970]: DEBUG nova.compute.manager [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1065.316863] env[61970]: DEBUG nova.virt.hardware [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1065.317024] env[61970]: DEBUG nova.virt.hardware [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1065.317194] env[61970]: DEBUG nova.virt.hardware [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1065.317631] env[61970]: DEBUG nova.virt.hardware [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1065.317726] env[61970]: DEBUG nova.virt.hardware [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1065.318022] env[61970]: DEBUG nova.virt.hardware [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1065.318080] env[61970]: DEBUG nova.virt.hardware [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1065.318215] env[61970]: DEBUG nova.virt.hardware [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1065.318386] env[61970]: DEBUG nova.virt.hardware [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1065.318555] env[61970]: DEBUG nova.virt.hardware [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1065.318727] env[61970]: DEBUG nova.virt.hardware [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1065.319618] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a4a0726-d4bb-40e5-9ad7-a7b1fc4eb998 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.327885] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0551a496-4c92-406f-8958-8b67b67d36ab {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.473333] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.206s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.473931] env[61970]: DEBUG nova.compute.manager [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1065.603805] env[61970]: DEBUG nova.compute.manager [req-f9f3712a-cf52-4b86-9d39-dec2815efb19 req-e51cfaf1-9079-49fa-ab17-c9d24b03ae81 service nova] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Received event network-vif-plugged-f89f6ca5-913c-4eff-ab7e-2199ee90ccd7 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1065.604185] env[61970]: DEBUG oslo_concurrency.lockutils [req-f9f3712a-cf52-4b86-9d39-dec2815efb19 req-e51cfaf1-9079-49fa-ab17-c9d24b03ae81 service nova] Acquiring lock "183133c6-7e4e-4dd4-b8e0-cb4f8338264c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.604463] env[61970]: DEBUG oslo_concurrency.lockutils [req-f9f3712a-cf52-4b86-9d39-dec2815efb19 req-e51cfaf1-9079-49fa-ab17-c9d24b03ae81 service nova] Lock "183133c6-7e4e-4dd4-b8e0-cb4f8338264c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.604628] env[61970]: DEBUG oslo_concurrency.lockutils [req-f9f3712a-cf52-4b86-9d39-dec2815efb19 req-e51cfaf1-9079-49fa-ab17-c9d24b03ae81 service nova] Lock "183133c6-7e4e-4dd4-b8e0-cb4f8338264c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.604799] env[61970]: DEBUG nova.compute.manager [req-f9f3712a-cf52-4b86-9d39-dec2815efb19 req-e51cfaf1-9079-49fa-ab17-c9d24b03ae81 service nova] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] No waiting events found dispatching network-vif-plugged-f89f6ca5-913c-4eff-ab7e-2199ee90ccd7 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1065.604968] env[61970]: WARNING nova.compute.manager [req-f9f3712a-cf52-4b86-9d39-dec2815efb19 req-e51cfaf1-9079-49fa-ab17-c9d24b03ae81 service nova] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Received unexpected event network-vif-plugged-f89f6ca5-913c-4eff-ab7e-2199ee90ccd7 for instance with vm_state building and task_state spawning. [ 1065.634217] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Releasing lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.634515] env[61970]: DEBUG nova.compute.manager [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Instance network_info: |[{"id": "2fa02dfe-2cb0-4a72-bc30-4f991339dcb9", "address": "fa:16:3e:df:19:94", "network": {"id": "70081995-3d6c-466e-b811-1c8e532e552f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-641102750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1d4724cbb0841859369af01f2c03db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aaf1b231-3660-4453-b4f3-44d825b9a5dd", "external-id": "nsx-vlan-transportzone-6", "segmentation_id": 6, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fa02dfe-2c", "ovs_interfaceid": "2fa02dfe-2cb0-4a72-bc30-4f991339dcb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1065.634964] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:df:19:94', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aaf1b231-3660-4453-b4f3-44d825b9a5dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2fa02dfe-2cb0-4a72-bc30-4f991339dcb9', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1065.642579] env[61970]: DEBUG oslo.service.loopingcall [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1065.643226] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1065.643493] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b269208e-2414-4dc0-b6e3-7b5de4232232 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.665093] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1065.665093] env[61970]: value = "task-1356148" [ 1065.665093] env[61970]: _type = "Task" [ 1065.665093] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.671237] env[61970]: DEBUG oslo_vmware.api [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356147, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.677106] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356148, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.685276] env[61970]: DEBUG nova.network.neutron [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Successfully updated port: f89f6ca5-913c-4eff-ab7e-2199ee90ccd7 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1065.754017] env[61970]: DEBUG oslo_concurrency.lockutils [None req-744a8596-242e-43b1-9148-23ed0d5c59d7 tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Lock "ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.693s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.872999] env[61970]: DEBUG nova.compute.manager [req-2538b6db-f71f-4a88-adf3-68d0660f6bdb req-52b767e6-02fc-4efc-a24a-cd52daa33f46 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Received event network-changed-2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1065.873232] env[61970]: DEBUG nova.compute.manager [req-2538b6db-f71f-4a88-adf3-68d0660f6bdb req-52b767e6-02fc-4efc-a24a-cd52daa33f46 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Refreshing instance network info cache due to event network-changed-2fa02dfe-2cb0-4a72-bc30-4f991339dcb9. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1065.873472] env[61970]: DEBUG oslo_concurrency.lockutils [req-2538b6db-f71f-4a88-adf3-68d0660f6bdb req-52b767e6-02fc-4efc-a24a-cd52daa33f46 service nova] Acquiring lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.873678] env[61970]: DEBUG oslo_concurrency.lockutils [req-2538b6db-f71f-4a88-adf3-68d0660f6bdb req-52b767e6-02fc-4efc-a24a-cd52daa33f46 service nova] Acquired lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.873855] env[61970]: DEBUG nova.network.neutron [req-2538b6db-f71f-4a88-adf3-68d0660f6bdb req-52b767e6-02fc-4efc-a24a-cd52daa33f46 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Refreshing network info cache for port 2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1065.905887] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Acquiring lock "ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.906115] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Lock "ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.906364] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Acquiring lock "ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.906588] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Lock "ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.906770] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Lock "ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.908964] env[61970]: INFO nova.compute.manager [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Terminating instance [ 1065.910992] env[61970]: DEBUG nova.compute.manager [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1065.911267] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1065.912171] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f27f6d-b822-46b8-9d04-09c38f7c7eab {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.919999] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1065.920274] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4285f409-368f-4ae8-a57a-db87e6d24080 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.927377] env[61970]: DEBUG oslo_vmware.api [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Waiting for the task: (returnval){ [ 1065.927377] env[61970]: value = "task-1356149" [ 1065.927377] env[61970]: _type = "Task" [ 1065.927377] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.939506] env[61970]: DEBUG oslo_vmware.api [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': task-1356149, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.980059] env[61970]: DEBUG nova.compute.utils [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1065.982070] env[61970]: DEBUG nova.compute.manager [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1065.982270] env[61970]: DEBUG nova.network.neutron [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1066.024145] env[61970]: DEBUG nova.policy [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2b8322c5de6483aaf032781d28e950d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09033ee688384ca287b4fdad6e67cb1f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 1066.172779] env[61970]: DEBUG oslo_vmware.api [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356147, 'name': ReconfigVM_Task, 'duration_secs': 1.177814} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.173761] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288874', 'volume_id': 'f1fea832-29d5-441d-a553-3c04a1903995', 'name': 'volume-f1fea832-29d5-441d-a553-3c04a1903995', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '31f9d37b-f4ff-4a8b-9477-8253f8305020', 'attached_at': '', 'detached_at': '', 'volume_id': 'f1fea832-29d5-441d-a553-3c04a1903995', 'serial': 'f1fea832-29d5-441d-a553-3c04a1903995'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1066.179548] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356148, 'name': CreateVM_Task, 'duration_secs': 0.394665} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.179764] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1066.180481] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.180890] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.181166] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1066.181485] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8615ba7-ee4c-4d5a-9734-d7a8f4c04f51 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.186483] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1066.186483] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526d5011-5313-2b29-1b8f-9d0c2011f6b5" [ 1066.186483] env[61970]: _type = "Task" [ 1066.186483] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.189999] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "refresh_cache-183133c6-7e4e-4dd4-b8e0-cb4f8338264c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.189999] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "refresh_cache-183133c6-7e4e-4dd4-b8e0-cb4f8338264c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.189999] env[61970]: DEBUG nova.network.neutron [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1066.196283] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526d5011-5313-2b29-1b8f-9d0c2011f6b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.353389] env[61970]: DEBUG nova.network.neutron [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Successfully created port: 696b0179-1272-4ae7-ab15-727fe4682c42 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1066.437220] env[61970]: DEBUG oslo_vmware.api [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': task-1356149, 'name': PowerOffVM_Task, 'duration_secs': 0.167005} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.437478] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1066.437735] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1066.437895] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2b08a73-d2ca-49e7-9e10-72fa2fcdd990 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.485235] env[61970]: DEBUG nova.compute.manager [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1066.518033] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1066.518381] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1066.518637] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Deleting the datastore file [datastore1] ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1066.519012] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c1362961-271e-4325-b233-2d51aceb50ba {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.525729] env[61970]: DEBUG oslo_vmware.api [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Waiting for the task: (returnval){ [ 1066.525729] env[61970]: value = "task-1356151" [ 1066.525729] env[61970]: _type = "Task" [ 1066.525729] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.533882] env[61970]: DEBUG oslo_vmware.api [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': task-1356151, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.570767] env[61970]: DEBUG nova.network.neutron [req-2538b6db-f71f-4a88-adf3-68d0660f6bdb req-52b767e6-02fc-4efc-a24a-cd52daa33f46 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Updated VIF entry in instance network info cache for port 2fa02dfe-2cb0-4a72-bc30-4f991339dcb9. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1066.571150] env[61970]: DEBUG nova.network.neutron [req-2538b6db-f71f-4a88-adf3-68d0660f6bdb req-52b767e6-02fc-4efc-a24a-cd52daa33f46 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Updating instance_info_cache with network_info: [{"id": "2fa02dfe-2cb0-4a72-bc30-4f991339dcb9", "address": "fa:16:3e:df:19:94", "network": {"id": "70081995-3d6c-466e-b811-1c8e532e552f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-641102750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1d4724cbb0841859369af01f2c03db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aaf1b231-3660-4453-b4f3-44d825b9a5dd", "external-id": "nsx-vlan-transportzone-6", "segmentation_id": 6, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fa02dfe-2c", "ovs_interfaceid": "2fa02dfe-2cb0-4a72-bc30-4f991339dcb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.697956] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526d5011-5313-2b29-1b8f-9d0c2011f6b5, 'name': SearchDatastore_Task, 'duration_secs': 0.009818} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.698702] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.698984] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1066.699291] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.699442] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.699627] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1066.699897] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9e67393a-9b03-4445-99d7-71a89c3f0238 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.707415] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1066.707775] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1066.708333] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2501a37-754a-4e77-8604-651c5e48e2e5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.713293] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1066.713293] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]527401b2-fbfa-e389-705c-e20bc64bd033" [ 1066.713293] env[61970]: _type = "Task" [ 1066.713293] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.722834] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]527401b2-fbfa-e389-705c-e20bc64bd033, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.724610] env[61970]: DEBUG nova.objects.instance [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lazy-loading 'flavor' on Instance uuid 31f9d37b-f4ff-4a8b-9477-8253f8305020 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.726327] env[61970]: DEBUG nova.network.neutron [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1066.895825] env[61970]: DEBUG nova.network.neutron [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Updating instance_info_cache with network_info: [{"id": "f89f6ca5-913c-4eff-ab7e-2199ee90ccd7", "address": "fa:16:3e:7a:43:45", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf89f6ca5-91", "ovs_interfaceid": "f89f6ca5-913c-4eff-ab7e-2199ee90ccd7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.035878] env[61970]: DEBUG oslo_vmware.api [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Task: {'id': task-1356151, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166659} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.036104] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1067.036293] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1067.036482] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1067.036751] env[61970]: INFO nova.compute.manager [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1067.036892] env[61970]: DEBUG oslo.service.loopingcall [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1067.037101] env[61970]: DEBUG nova.compute.manager [-] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1067.037199] env[61970]: DEBUG nova.network.neutron [-] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1067.073562] env[61970]: DEBUG oslo_concurrency.lockutils [req-2538b6db-f71f-4a88-adf3-68d0660f6bdb req-52b767e6-02fc-4efc-a24a-cd52daa33f46 service nova] Releasing lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.223980] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]527401b2-fbfa-e389-705c-e20bc64bd033, 'name': SearchDatastore_Task, 'duration_secs': 0.008675} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.224733] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b670378f-8680-40fe-83de-07e57a28af6a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.234021] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1067.234021] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5210383a-5e4c-cb8c-9e36-22f378de88c8" [ 1067.234021] env[61970]: _type = "Task" [ 1067.234021] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.241618] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5210383a-5e4c-cb8c-9e36-22f378de88c8, 'name': SearchDatastore_Task} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.241857] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.242124] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6/3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1067.242432] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88a0fc4c-0810-4166-9098-3145fa1254e4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.250830] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1067.250830] env[61970]: value = "task-1356152" [ 1067.250830] env[61970]: _type = "Task" [ 1067.250830] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.258574] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356152, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.398348] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "refresh_cache-183133c6-7e4e-4dd4-b8e0-cb4f8338264c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.398687] env[61970]: DEBUG nova.compute.manager [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Instance network_info: |[{"id": "f89f6ca5-913c-4eff-ab7e-2199ee90ccd7", "address": "fa:16:3e:7a:43:45", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf89f6ca5-91", "ovs_interfaceid": "f89f6ca5-913c-4eff-ab7e-2199ee90ccd7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1067.399933] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7a:43:45', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b356db78-99c7-4464-822c-fc7e193f7878', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f89f6ca5-913c-4eff-ab7e-2199ee90ccd7', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1067.407617] env[61970]: DEBUG oslo.service.loopingcall [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1067.407851] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1067.408215] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-add7ede2-8ada-490a-b20b-7b0bec8e6b95 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.430814] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1067.430814] env[61970]: value = "task-1356153" [ 1067.430814] env[61970]: _type = "Task" [ 1067.430814] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.438906] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356153, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.496163] env[61970]: DEBUG nova.compute.manager [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1067.525706] env[61970]: DEBUG nova.virt.hardware [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1067.525986] env[61970]: DEBUG nova.virt.hardware [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1067.526172] env[61970]: DEBUG nova.virt.hardware [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1067.526364] env[61970]: DEBUG nova.virt.hardware [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1067.526519] env[61970]: DEBUG nova.virt.hardware [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1067.526670] env[61970]: DEBUG nova.virt.hardware [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1067.526909] env[61970]: DEBUG nova.virt.hardware [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1067.527063] env[61970]: DEBUG nova.virt.hardware [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1067.527246] env[61970]: DEBUG nova.virt.hardware [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1067.527418] env[61970]: DEBUG nova.virt.hardware [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1067.527592] env[61970]: DEBUG nova.virt.hardware [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1067.528520] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90d4c81-8c79-4a03-904a-d234533f0cc0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.537276] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3b17d3-1e2a-48c3-be8f-093bf6cf973a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.632337] env[61970]: DEBUG nova.compute.manager [req-ac40fb86-06f2-4a00-8941-020197242e22 req-1b09619b-c1af-42ba-af7f-3e85e024cbef service nova] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Received event network-changed-f89f6ca5-913c-4eff-ab7e-2199ee90ccd7 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1067.632649] env[61970]: DEBUG nova.compute.manager [req-ac40fb86-06f2-4a00-8941-020197242e22 req-1b09619b-c1af-42ba-af7f-3e85e024cbef service nova] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Refreshing instance network info cache due to event network-changed-f89f6ca5-913c-4eff-ab7e-2199ee90ccd7. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1067.632791] env[61970]: DEBUG oslo_concurrency.lockutils [req-ac40fb86-06f2-4a00-8941-020197242e22 req-1b09619b-c1af-42ba-af7f-3e85e024cbef service nova] Acquiring lock "refresh_cache-183133c6-7e4e-4dd4-b8e0-cb4f8338264c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.632940] env[61970]: DEBUG oslo_concurrency.lockutils [req-ac40fb86-06f2-4a00-8941-020197242e22 req-1b09619b-c1af-42ba-af7f-3e85e024cbef service nova] Acquired lock "refresh_cache-183133c6-7e4e-4dd4-b8e0-cb4f8338264c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.633124] env[61970]: DEBUG nova.network.neutron [req-ac40fb86-06f2-4a00-8941-020197242e22 req-1b09619b-c1af-42ba-af7f-3e85e024cbef service nova] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Refreshing network info cache for port f89f6ca5-913c-4eff-ab7e-2199ee90ccd7 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1067.731771] env[61970]: DEBUG nova.network.neutron [-] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.736785] env[61970]: DEBUG oslo_concurrency.lockutils [None req-aa3026b7-ecd5-4a28-b3d7-2f63ed6fa0db tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "31f9d37b-f4ff-4a8b-9477-8253f8305020" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.242s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.765397] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356152, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.880870] env[61970]: DEBUG nova.network.neutron [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Successfully updated port: 696b0179-1272-4ae7-ab15-727fe4682c42 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1067.898604] env[61970]: DEBUG nova.compute.manager [req-99ea3f53-bf6a-4860-b3df-a19437bd8a9d req-cb9384c9-e51a-4d4c-84ff-ef9cab472c71 service nova] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Received event network-vif-deleted-572f62fb-dd55-4cd7-8c2f-a78ef73eb69f {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1067.940508] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356153, 'name': CreateVM_Task} progress is 25%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.235128] env[61970]: INFO nova.compute.manager [-] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Took 1.20 seconds to deallocate network for instance. [ 1068.262408] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356152, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56696} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.262686] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6/3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1068.262907] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1068.263656] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-864317bb-e997-4d1f-87f9-8a2f3aeecdc3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.269886] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1068.269886] env[61970]: value = "task-1356154" [ 1068.269886] env[61970]: _type = "Task" [ 1068.269886] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.281102] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356154, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.362955] env[61970]: DEBUG nova.network.neutron [req-ac40fb86-06f2-4a00-8941-020197242e22 req-1b09619b-c1af-42ba-af7f-3e85e024cbef service nova] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Updated VIF entry in instance network info cache for port f89f6ca5-913c-4eff-ab7e-2199ee90ccd7. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1068.363349] env[61970]: DEBUG nova.network.neutron [req-ac40fb86-06f2-4a00-8941-020197242e22 req-1b09619b-c1af-42ba-af7f-3e85e024cbef service nova] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Updating instance_info_cache with network_info: [{"id": "f89f6ca5-913c-4eff-ab7e-2199ee90ccd7", "address": "fa:16:3e:7a:43:45", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf89f6ca5-91", "ovs_interfaceid": "f89f6ca5-913c-4eff-ab7e-2199ee90ccd7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.382817] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.382948] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.383100] env[61970]: DEBUG nova.network.neutron [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1068.440977] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356153, 'name': CreateVM_Task, 'duration_secs': 0.60442} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.441281] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1068.441967] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.442160] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.442513] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1068.442771] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f07b4b5-7d63-4384-8d1e-2b14c30cf00f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.447028] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1068.447028] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52766b1b-472a-3690-fc4f-0ff9b31e4881" [ 1068.447028] env[61970]: _type = "Task" [ 1068.447028] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.454495] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52766b1b-472a-3690-fc4f-0ff9b31e4881, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.745144] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.745481] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.745745] env[61970]: DEBUG nova.objects.instance [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Lazy-loading 'resources' on Instance uuid ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.783040] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356154, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069139} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.783040] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1068.783040] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e624c385-09b3-4a16-b769-a5ff3e9c4313 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.798153] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "31f9d37b-f4ff-4a8b-9477-8253f8305020" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.798378] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "31f9d37b-f4ff-4a8b-9477-8253f8305020" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.798608] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "31f9d37b-f4ff-4a8b-9477-8253f8305020-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.798768] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "31f9d37b-f4ff-4a8b-9477-8253f8305020-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.798941] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "31f9d37b-f4ff-4a8b-9477-8253f8305020-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.808461] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6/3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1068.808915] env[61970]: INFO nova.compute.manager [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Terminating instance [ 1068.810364] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-673f2826-cdfc-488e-962f-d57cc9345ac3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.824323] env[61970]: DEBUG nova.compute.manager [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1068.824522] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1068.825216] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-866903ae-cf97-4650-a74a-6a0e85c966b4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.832445] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1068.832665] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c3c136e-b89b-4e66-ad98-b27155b5efc7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.834812] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1068.834812] env[61970]: value = "task-1356155" [ 1068.834812] env[61970]: _type = "Task" [ 1068.834812] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.839723] env[61970]: DEBUG oslo_vmware.api [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1068.839723] env[61970]: value = "task-1356156" [ 1068.839723] env[61970]: _type = "Task" [ 1068.839723] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.842702] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356155, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.850441] env[61970]: DEBUG oslo_vmware.api [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356156, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.866127] env[61970]: DEBUG oslo_concurrency.lockutils [req-ac40fb86-06f2-4a00-8941-020197242e22 req-1b09619b-c1af-42ba-af7f-3e85e024cbef service nova] Releasing lock "refresh_cache-183133c6-7e4e-4dd4-b8e0-cb4f8338264c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.914050] env[61970]: DEBUG nova.network.neutron [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1068.960206] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52766b1b-472a-3690-fc4f-0ff9b31e4881, 'name': SearchDatastore_Task, 'duration_secs': 0.009219} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.960523] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.960762] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1068.961020] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.961233] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.961445] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1068.961714] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-556a6126-d846-4696-8e37-cf506f8c6e12 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.970387] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1068.970569] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1068.971534] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-423c622c-79de-4100-98e8-1b6a86e59973 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.976573] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1068.976573] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5299b706-31f0-b735-9055-c7a4063384e9" [ 1068.976573] env[61970]: _type = "Task" [ 1068.976573] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.985965] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5299b706-31f0-b735-9055-c7a4063384e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.045895] env[61970]: DEBUG nova.network.neutron [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating instance_info_cache with network_info: [{"id": "696b0179-1272-4ae7-ab15-727fe4682c42", "address": "fa:16:3e:9e:3f:74", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap696b0179-12", "ovs_interfaceid": "696b0179-1272-4ae7-ab15-727fe4682c42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.334686] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-403f987c-b1cf-4d12-9210-9dfa9a415083 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.349292] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae833f5-a805-450b-b204-dd1a44fe99bd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.352392] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356155, 'name': ReconfigVM_Task, 'duration_secs': 0.425919} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.352850] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6/3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1069.354275] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f90cd76c-1e2f-472d-8261-28876ebcb630 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.381602] env[61970]: DEBUG oslo_vmware.api [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356156, 'name': PowerOffVM_Task, 'duration_secs': 0.224635} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.382459] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1069.382691] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1069.383408] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d23478c4-5c3b-46fa-af0d-7795f2616857 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.385719] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f417419f-935b-41a5-82c6-db7e296f87e3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.388276] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1069.388276] env[61970]: value = "task-1356157" [ 1069.388276] env[61970]: _type = "Task" [ 1069.388276] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.395750] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40e5d741-ded5-4ccb-9c8c-b51fee8c0fea {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.402044] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356157, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.412327] env[61970]: DEBUG nova.compute.provider_tree [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1069.460653] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1069.460893] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1069.461098] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Deleting the datastore file [datastore2] 31f9d37b-f4ff-4a8b-9477-8253f8305020 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1069.461431] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b0c4fdea-5cf0-4db4-b4c7-655eac43b638 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.467795] env[61970]: DEBUG oslo_vmware.api [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1069.467795] env[61970]: value = "task-1356159" [ 1069.467795] env[61970]: _type = "Task" [ 1069.467795] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.475531] env[61970]: DEBUG oslo_vmware.api [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356159, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.485382] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5299b706-31f0-b735-9055-c7a4063384e9, 'name': SearchDatastore_Task, 'duration_secs': 0.018934} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.486414] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e65a681d-7cdd-45a5-b263-40762770b959 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.491017] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1069.491017] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5277db40-ca26-cb74-ff13-dff03fdf230f" [ 1069.491017] env[61970]: _type = "Task" [ 1069.491017] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.498102] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5277db40-ca26-cb74-ff13-dff03fdf230f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.549144] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.549518] env[61970]: DEBUG nova.compute.manager [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Instance network_info: |[{"id": "696b0179-1272-4ae7-ab15-727fe4682c42", "address": "fa:16:3e:9e:3f:74", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap696b0179-12", "ovs_interfaceid": "696b0179-1272-4ae7-ab15-727fe4682c42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1069.550024] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9e:3f:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '298bb8ef-4765-494c-b157-7a349218bd1e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '696b0179-1272-4ae7-ab15-727fe4682c42', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1069.558031] env[61970]: DEBUG oslo.service.loopingcall [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.558644] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1069.558915] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9d466f83-32ec-40c8-b9b9-d9f4db5631e8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.577274] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1069.577274] env[61970]: value = "task-1356160" [ 1069.577274] env[61970]: _type = "Task" [ 1069.577274] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.584309] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356160, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.658412] env[61970]: DEBUG nova.compute.manager [req-7c9ffc48-20e1-46ba-b519-9e6dcf9329e2 req-23fe1a29-55ec-4221-a987-e8cf3afa1eab service nova] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Received event network-vif-plugged-696b0179-1272-4ae7-ab15-727fe4682c42 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1069.658641] env[61970]: DEBUG oslo_concurrency.lockutils [req-7c9ffc48-20e1-46ba-b519-9e6dcf9329e2 req-23fe1a29-55ec-4221-a987-e8cf3afa1eab service nova] Acquiring lock "5a8cdb36-1845-4806-a6bc-22afb47886e2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.658854] env[61970]: DEBUG oslo_concurrency.lockutils [req-7c9ffc48-20e1-46ba-b519-9e6dcf9329e2 req-23fe1a29-55ec-4221-a987-e8cf3afa1eab service nova] Lock "5a8cdb36-1845-4806-a6bc-22afb47886e2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.659040] env[61970]: DEBUG oslo_concurrency.lockutils [req-7c9ffc48-20e1-46ba-b519-9e6dcf9329e2 req-23fe1a29-55ec-4221-a987-e8cf3afa1eab service nova] Lock "5a8cdb36-1845-4806-a6bc-22afb47886e2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.659382] env[61970]: DEBUG nova.compute.manager [req-7c9ffc48-20e1-46ba-b519-9e6dcf9329e2 req-23fe1a29-55ec-4221-a987-e8cf3afa1eab service nova] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] No waiting events found dispatching network-vif-plugged-696b0179-1272-4ae7-ab15-727fe4682c42 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1069.659622] env[61970]: WARNING nova.compute.manager [req-7c9ffc48-20e1-46ba-b519-9e6dcf9329e2 req-23fe1a29-55ec-4221-a987-e8cf3afa1eab service nova] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Received unexpected event network-vif-plugged-696b0179-1272-4ae7-ab15-727fe4682c42 for instance with vm_state building and task_state spawning. [ 1069.659802] env[61970]: DEBUG nova.compute.manager [req-7c9ffc48-20e1-46ba-b519-9e6dcf9329e2 req-23fe1a29-55ec-4221-a987-e8cf3afa1eab service nova] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Received event network-changed-696b0179-1272-4ae7-ab15-727fe4682c42 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1069.660060] env[61970]: DEBUG nova.compute.manager [req-7c9ffc48-20e1-46ba-b519-9e6dcf9329e2 req-23fe1a29-55ec-4221-a987-e8cf3afa1eab service nova] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Refreshing instance network info cache due to event network-changed-696b0179-1272-4ae7-ab15-727fe4682c42. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1069.660297] env[61970]: DEBUG oslo_concurrency.lockutils [req-7c9ffc48-20e1-46ba-b519-9e6dcf9329e2 req-23fe1a29-55ec-4221-a987-e8cf3afa1eab service nova] Acquiring lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.660448] env[61970]: DEBUG oslo_concurrency.lockutils [req-7c9ffc48-20e1-46ba-b519-9e6dcf9329e2 req-23fe1a29-55ec-4221-a987-e8cf3afa1eab service nova] Acquired lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.660612] env[61970]: DEBUG nova.network.neutron [req-7c9ffc48-20e1-46ba-b519-9e6dcf9329e2 req-23fe1a29-55ec-4221-a987-e8cf3afa1eab service nova] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Refreshing network info cache for port 696b0179-1272-4ae7-ab15-727fe4682c42 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1069.898979] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356157, 'name': Rename_Task, 'duration_secs': 0.157894} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.899347] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1069.899609] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b465fa6e-84f9-4766-baef-9a81898f7e47 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.905763] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1069.905763] env[61970]: value = "task-1356161" [ 1069.905763] env[61970]: _type = "Task" [ 1069.905763] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.914310] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356161, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.932912] env[61970]: ERROR nova.scheduler.client.report [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] [req-c03309b8-b98a-42cd-93fd-3d2bb9cc946a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fc0f3e4e-5ef5-45e9-8335-cc68743405aa. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c03309b8-b98a-42cd-93fd-3d2bb9cc946a"}]} [ 1069.949168] env[61970]: DEBUG nova.scheduler.client.report [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Refreshing inventories for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1069.963025] env[61970]: DEBUG nova.scheduler.client.report [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Updating ProviderTree inventory for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1069.963399] env[61970]: DEBUG nova.compute.provider_tree [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1069.977366] env[61970]: DEBUG oslo_vmware.api [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356159, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155096} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.978237] env[61970]: DEBUG nova.scheduler.client.report [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Refreshing aggregate associations for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa, aggregates: None {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1069.980063] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1069.980269] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1069.980453] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1069.980631] env[61970]: INFO nova.compute.manager [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1069.980876] env[61970]: DEBUG oslo.service.loopingcall [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.982023] env[61970]: DEBUG nova.compute.manager [-] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1069.982023] env[61970]: DEBUG nova.network.neutron [-] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1069.997669] env[61970]: DEBUG nova.scheduler.client.report [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Refreshing trait associations for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1070.004820] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5277db40-ca26-cb74-ff13-dff03fdf230f, 'name': SearchDatastore_Task, 'duration_secs': 0.020123} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.005135] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.005405] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 183133c6-7e4e-4dd4-b8e0-cb4f8338264c/183133c6-7e4e-4dd4-b8e0-cb4f8338264c.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1070.005706] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4820cedf-0ebd-4a79-b53c-6818d575e520 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.012039] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1070.012039] env[61970]: value = "task-1356162" [ 1070.012039] env[61970]: _type = "Task" [ 1070.012039] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.021935] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356162, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.089823] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356160, 'name': CreateVM_Task, 'duration_secs': 0.349175} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.090011] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1070.090790] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a7561ed-4165-4f17-ac6f-611663a284e4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.093730] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.093944] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.094329] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1070.094650] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c7c8391-9347-40ac-b004-d2a8d62dca4f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.102124] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e4f9e29-0a77-4819-a131-30a542971b50 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.105223] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1070.105223] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ad856e-db7e-574f-f934-414bf0353d7f" [ 1070.105223] env[61970]: _type = "Task" [ 1070.105223] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.133124] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec03e55f-f399-4c52-a039-48b57a858268 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.139082] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ad856e-db7e-574f-f934-414bf0353d7f, 'name': SearchDatastore_Task, 'duration_secs': 0.009472} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.139728] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.139995] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1070.140249] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.140404] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.140584] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1070.140835] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3802bc9b-d1a4-4f7d-b136-754da82dea70 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.145998] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf15a6a1-c9a0-4dcd-87dc-b355bc98f886 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.150652] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1070.150846] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1070.151909] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a5dc484-e060-488d-90a5-af3b822f532b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.161906] env[61970]: DEBUG nova.compute.provider_tree [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1070.166800] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1070.166800] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]520f3438-5fca-cf6d-f336-eeae9a972efe" [ 1070.166800] env[61970]: _type = "Task" [ 1070.166800] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.174581] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]520f3438-5fca-cf6d-f336-eeae9a972efe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.416545] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356161, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.419181] env[61970]: DEBUG nova.network.neutron [req-7c9ffc48-20e1-46ba-b519-9e6dcf9329e2 req-23fe1a29-55ec-4221-a987-e8cf3afa1eab service nova] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updated VIF entry in instance network info cache for port 696b0179-1272-4ae7-ab15-727fe4682c42. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1070.419181] env[61970]: DEBUG nova.network.neutron [req-7c9ffc48-20e1-46ba-b519-9e6dcf9329e2 req-23fe1a29-55ec-4221-a987-e8cf3afa1eab service nova] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating instance_info_cache with network_info: [{"id": "696b0179-1272-4ae7-ab15-727fe4682c42", "address": "fa:16:3e:9e:3f:74", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap696b0179-12", "ovs_interfaceid": "696b0179-1272-4ae7-ab15-727fe4682c42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.524378] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356162, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.682440] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]520f3438-5fca-cf6d-f336-eeae9a972efe, 'name': SearchDatastore_Task, 'duration_secs': 0.008825} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.683331] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0cf7ac4e-0f11-42eb-9665-a20118550726 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.688503] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1070.688503] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f72ce1-a9b8-5b99-5f6f-097a3c060016" [ 1070.688503] env[61970]: _type = "Task" [ 1070.688503] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.697348] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f72ce1-a9b8-5b99-5f6f-097a3c060016, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.698359] env[61970]: DEBUG nova.scheduler.client.report [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Updated inventory for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with generation 131 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1070.698665] env[61970]: DEBUG nova.compute.provider_tree [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Updating resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa generation from 131 to 132 during operation: update_inventory {{(pid=61970) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1070.698903] env[61970]: DEBUG nova.compute.provider_tree [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1070.916976] env[61970]: DEBUG oslo_vmware.api [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356161, 'name': PowerOnVM_Task, 'duration_secs': 0.814697} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.917310] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1070.917543] env[61970]: INFO nova.compute.manager [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Took 7.85 seconds to spawn the instance on the hypervisor. [ 1070.917793] env[61970]: DEBUG nova.compute.manager [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1070.918569] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-178cb706-74ce-414e-97b0-e5652d3eda50 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.921646] env[61970]: DEBUG oslo_concurrency.lockutils [req-7c9ffc48-20e1-46ba-b519-9e6dcf9329e2 req-23fe1a29-55ec-4221-a987-e8cf3afa1eab service nova] Releasing lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.000988] env[61970]: DEBUG nova.network.neutron [-] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.024556] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356162, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531064} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.024817] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 183133c6-7e4e-4dd4-b8e0-cb4f8338264c/183133c6-7e4e-4dd4-b8e0-cb4f8338264c.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1071.025050] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1071.025298] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2cb7ae9c-8500-41b7-94f7-e0e47c5be48e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.032158] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1071.032158] env[61970]: value = "task-1356163" [ 1071.032158] env[61970]: _type = "Task" [ 1071.032158] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.039642] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356163, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.198451] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f72ce1-a9b8-5b99-5f6f-097a3c060016, 'name': SearchDatastore_Task, 'duration_secs': 0.010269} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.198721] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.198973] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 5a8cdb36-1845-4806-a6bc-22afb47886e2/5a8cdb36-1845-4806-a6bc-22afb47886e2.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1071.199260] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cde2dbc5-cf2d-4989-8e35-3558993222d6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.203988] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.459s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.208027] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1071.208027] env[61970]: value = "task-1356164" [ 1071.208027] env[61970]: _type = "Task" [ 1071.208027] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.215937] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356164, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.223294] env[61970]: INFO nova.scheduler.client.report [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Deleted allocations for instance ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97 [ 1071.436513] env[61970]: INFO nova.compute.manager [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Took 12.60 seconds to build instance. [ 1071.502533] env[61970]: INFO nova.compute.manager [-] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Took 1.52 seconds to deallocate network for instance. [ 1071.543776] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356163, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0629} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.544020] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1071.544745] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6600c08a-99e9-4a84-a0b6-222473bb72ca {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.567446] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 183133c6-7e4e-4dd4-b8e0-cb4f8338264c/183133c6-7e4e-4dd4-b8e0-cb4f8338264c.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1071.567721] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b8e9459-535e-44b2-ac7a-48ddcd336a87 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.587199] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1071.587199] env[61970]: value = "task-1356165" [ 1071.587199] env[61970]: _type = "Task" [ 1071.587199] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.594929] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356165, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.683200] env[61970]: DEBUG nova.compute.manager [req-f910882d-a4ae-41bc-ae7c-87cb8dc359c5 req-31a816b9-cb4a-4aa9-b7cd-e96865d1190c service nova] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Received event network-vif-deleted-b39ddbde-1199-4f98-9fe2-cd05d19bc96e {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1071.717636] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356164, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.730569] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8eab18b3-e1d9-42eb-98e5-1ab74b5aac9a tempest-InstanceActionsV221TestJSON-1607068723 tempest-InstanceActionsV221TestJSON-1607068723-project-member] Lock "ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.824s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.939680] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b0c8f185-5577-461a-85ed-adbe92eb6486 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.113s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.008571] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.008855] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.009105] env[61970]: DEBUG nova.objects.instance [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lazy-loading 'resources' on Instance uuid 31f9d37b-f4ff-4a8b-9477-8253f8305020 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.098306] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356165, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.218283] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356164, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.596032] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e68bfeec-9a1c-4d08-89cd-6f573edcae50 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.604209] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356165, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.606837] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c9fb0a8-9a1b-4d9f-af65-c057ceead6af {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.646102] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e56a44d-2c3b-47f3-8fef-cf087c72eca9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.652589] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0da8c76-e3ac-4a2b-9e68-c00bc0332b2d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.666074] env[61970]: DEBUG nova.compute.provider_tree [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1072.718385] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356164, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.18434} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.718598] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 5a8cdb36-1845-4806-a6bc-22afb47886e2/5a8cdb36-1845-4806-a6bc-22afb47886e2.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1072.718815] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1072.719084] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b0a0553-77eb-4589-8ac6-37cd29ba4a2d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.725419] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1072.725419] env[61970]: value = "task-1356166" [ 1072.725419] env[61970]: _type = "Task" [ 1072.725419] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.733927] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356166, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.836887] env[61970]: DEBUG nova.compute.manager [req-2952da91-7dab-4e6f-98b2-95551c532162 req-1e9fbd26-4266-4123-997f-caea5cd416c1 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Received event network-changed-2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1072.838022] env[61970]: DEBUG nova.compute.manager [req-2952da91-7dab-4e6f-98b2-95551c532162 req-1e9fbd26-4266-4123-997f-caea5cd416c1 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Refreshing instance network info cache due to event network-changed-2fa02dfe-2cb0-4a72-bc30-4f991339dcb9. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1072.838022] env[61970]: DEBUG oslo_concurrency.lockutils [req-2952da91-7dab-4e6f-98b2-95551c532162 req-1e9fbd26-4266-4123-997f-caea5cd416c1 service nova] Acquiring lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.838022] env[61970]: DEBUG oslo_concurrency.lockutils [req-2952da91-7dab-4e6f-98b2-95551c532162 req-1e9fbd26-4266-4123-997f-caea5cd416c1 service nova] Acquired lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.838022] env[61970]: DEBUG nova.network.neutron [req-2952da91-7dab-4e6f-98b2-95551c532162 req-1e9fbd26-4266-4123-997f-caea5cd416c1 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Refreshing network info cache for port 2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1073.099831] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356165, 'name': ReconfigVM_Task, 'duration_secs': 1.198843} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.100330] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 183133c6-7e4e-4dd4-b8e0-cb4f8338264c/183133c6-7e4e-4dd4-b8e0-cb4f8338264c.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1073.101156] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cb9e9bce-499c-4d31-a6f1-1cef4be42ba7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.111678] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1073.111678] env[61970]: value = "task-1356167" [ 1073.111678] env[61970]: _type = "Task" [ 1073.111678] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.119868] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356167, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.201077] env[61970]: DEBUG nova.scheduler.client.report [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Updated inventory for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with generation 132 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1073.201756] env[61970]: DEBUG nova.compute.provider_tree [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Updating resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa generation from 132 to 133 during operation: update_inventory {{(pid=61970) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1073.202115] env[61970]: DEBUG nova.compute.provider_tree [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1073.239130] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356166, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062617} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.239495] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1073.240309] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c14220-533c-4e36-8f38-e6749f404160 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.263949] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 5a8cdb36-1845-4806-a6bc-22afb47886e2/5a8cdb36-1845-4806-a6bc-22afb47886e2.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1073.264608] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12c443e3-d3a7-43af-a91c-74610adc48d0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.284281] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1073.284281] env[61970]: value = "task-1356168" [ 1073.284281] env[61970]: _type = "Task" [ 1073.284281] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.293585] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356168, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.589903] env[61970]: DEBUG nova.network.neutron [req-2952da91-7dab-4e6f-98b2-95551c532162 req-1e9fbd26-4266-4123-997f-caea5cd416c1 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Updated VIF entry in instance network info cache for port 2fa02dfe-2cb0-4a72-bc30-4f991339dcb9. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1073.591061] env[61970]: DEBUG nova.network.neutron [req-2952da91-7dab-4e6f-98b2-95551c532162 req-1e9fbd26-4266-4123-997f-caea5cd416c1 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Updating instance_info_cache with network_info: [{"id": "2fa02dfe-2cb0-4a72-bc30-4f991339dcb9", "address": "fa:16:3e:df:19:94", "network": {"id": "70081995-3d6c-466e-b811-1c8e532e552f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-641102750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1d4724cbb0841859369af01f2c03db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aaf1b231-3660-4453-b4f3-44d825b9a5dd", "external-id": "nsx-vlan-transportzone-6", "segmentation_id": 6, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fa02dfe-2c", "ovs_interfaceid": "2fa02dfe-2cb0-4a72-bc30-4f991339dcb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.620388] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356167, 'name': Rename_Task, 'duration_secs': 0.232802} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.620738] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1073.621060] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-53a5f9c7-8c44-42ef-bff0-d7475d620360 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.627754] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1073.627754] env[61970]: value = "task-1356169" [ 1073.627754] env[61970]: _type = "Task" [ 1073.627754] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.636753] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356169, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.707283] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.698s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.731928] env[61970]: INFO nova.scheduler.client.report [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Deleted allocations for instance 31f9d37b-f4ff-4a8b-9477-8253f8305020 [ 1073.794625] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356168, 'name': ReconfigVM_Task, 'duration_secs': 0.415181} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.794915] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 5a8cdb36-1845-4806-a6bc-22afb47886e2/5a8cdb36-1845-4806-a6bc-22afb47886e2.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1073.795643] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-175f73e2-aa3a-4a8c-8b12-7e82bc8bb9b2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.801801] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1073.801801] env[61970]: value = "task-1356170" [ 1073.801801] env[61970]: _type = "Task" [ 1073.801801] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.810051] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356170, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.094631] env[61970]: DEBUG oslo_concurrency.lockutils [req-2952da91-7dab-4e6f-98b2-95551c532162 req-1e9fbd26-4266-4123-997f-caea5cd416c1 service nova] Releasing lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.140247] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356169, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.239636] env[61970]: DEBUG oslo_concurrency.lockutils [None req-bff4cf69-e33d-44fa-94dc-ec2d90ebc00b tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "31f9d37b-f4ff-4a8b-9477-8253f8305020" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.441s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.315876] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356170, 'name': Rename_Task, 'duration_secs': 0.400434} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.316316] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1074.316663] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f6009fb8-21dd-4d8f-a022-9d928a20c0e2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.323949] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1074.323949] env[61970]: value = "task-1356171" [ 1074.323949] env[61970]: _type = "Task" [ 1074.323949] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.332513] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356171, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.639467] env[61970]: DEBUG oslo_vmware.api [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356169, 'name': PowerOnVM_Task, 'duration_secs': 0.860854} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.639861] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1074.639924] env[61970]: INFO nova.compute.manager [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Took 9.35 seconds to spawn the instance on the hypervisor. [ 1074.640108] env[61970]: DEBUG nova.compute.manager [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1074.640863] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e49ad31-7341-4311-8a61-741c36e94fac {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.834168] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356171, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.160245] env[61970]: INFO nova.compute.manager [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Took 14.36 seconds to build instance. [ 1075.173268] env[61970]: DEBUG nova.compute.manager [req-220e6c26-1943-4c45-a6ce-68bfd29dfb1a req-3ea2bad1-094e-40cc-98c5-43e264d40b8b service nova] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Received event network-changed-f89f6ca5-913c-4eff-ab7e-2199ee90ccd7 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1075.173545] env[61970]: DEBUG nova.compute.manager [req-220e6c26-1943-4c45-a6ce-68bfd29dfb1a req-3ea2bad1-094e-40cc-98c5-43e264d40b8b service nova] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Refreshing instance network info cache due to event network-changed-f89f6ca5-913c-4eff-ab7e-2199ee90ccd7. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1075.173780] env[61970]: DEBUG oslo_concurrency.lockutils [req-220e6c26-1943-4c45-a6ce-68bfd29dfb1a req-3ea2bad1-094e-40cc-98c5-43e264d40b8b service nova] Acquiring lock "refresh_cache-183133c6-7e4e-4dd4-b8e0-cb4f8338264c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.173931] env[61970]: DEBUG oslo_concurrency.lockutils [req-220e6c26-1943-4c45-a6ce-68bfd29dfb1a req-3ea2bad1-094e-40cc-98c5-43e264d40b8b service nova] Acquired lock "refresh_cache-183133c6-7e4e-4dd4-b8e0-cb4f8338264c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.174542] env[61970]: DEBUG nova.network.neutron [req-220e6c26-1943-4c45-a6ce-68bfd29dfb1a req-3ea2bad1-094e-40cc-98c5-43e264d40b8b service nova] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Refreshing network info cache for port f89f6ca5-913c-4eff-ab7e-2199ee90ccd7 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1075.336504] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356171, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.663718] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c7e04d20-62e9-4d82-a923-76a6da4a0967 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "183133c6-7e4e-4dd4-b8e0-cb4f8338264c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.874s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.836621] env[61970]: DEBUG oslo_vmware.api [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356171, 'name': PowerOnVM_Task, 'duration_secs': 1.390164} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.836886] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1075.837100] env[61970]: INFO nova.compute.manager [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Took 8.34 seconds to spawn the instance on the hypervisor. [ 1075.837279] env[61970]: DEBUG nova.compute.manager [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1075.838081] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3123070c-31d2-4221-a199-69c798944cc4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.877761] env[61970]: DEBUG nova.network.neutron [req-220e6c26-1943-4c45-a6ce-68bfd29dfb1a req-3ea2bad1-094e-40cc-98c5-43e264d40b8b service nova] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Updated VIF entry in instance network info cache for port f89f6ca5-913c-4eff-ab7e-2199ee90ccd7. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1075.877761] env[61970]: DEBUG nova.network.neutron [req-220e6c26-1943-4c45-a6ce-68bfd29dfb1a req-3ea2bad1-094e-40cc-98c5-43e264d40b8b service nova] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Updating instance_info_cache with network_info: [{"id": "f89f6ca5-913c-4eff-ab7e-2199ee90ccd7", "address": "fa:16:3e:7a:43:45", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf89f6ca5-91", "ovs_interfaceid": "f89f6ca5-913c-4eff-ab7e-2199ee90ccd7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.890593] env[61970]: DEBUG oslo_concurrency.lockutils [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "df0a2338-07e0-4148-8729-a290b88207b3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.890821] env[61970]: DEBUG oslo_concurrency.lockutils [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "df0a2338-07e0-4148-8729-a290b88207b3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.358514] env[61970]: INFO nova.compute.manager [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Took 14.79 seconds to build instance. [ 1076.381512] env[61970]: DEBUG oslo_concurrency.lockutils [req-220e6c26-1943-4c45-a6ce-68bfd29dfb1a req-3ea2bad1-094e-40cc-98c5-43e264d40b8b service nova] Releasing lock "refresh_cache-183133c6-7e4e-4dd4-b8e0-cb4f8338264c" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.393880] env[61970]: DEBUG nova.compute.manager [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1076.860295] env[61970]: DEBUG oslo_concurrency.lockutils [None req-0b454429-a35d-429c-8656-e9eadbbe8062 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "5a8cdb36-1845-4806-a6bc-22afb47886e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.306s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.916922] env[61970]: DEBUG oslo_concurrency.lockutils [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.917246] env[61970]: DEBUG oslo_concurrency.lockutils [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.918776] env[61970]: INFO nova.compute.claims [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1077.200151] env[61970]: DEBUG nova.compute.manager [req-6e0fefe9-90c8-4cfa-bbe7-610effd92af5 req-a8713f6c-b61c-41e8-a6c5-0a26bc4f7da2 service nova] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Received event network-changed-696b0179-1272-4ae7-ab15-727fe4682c42 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1077.200151] env[61970]: DEBUG nova.compute.manager [req-6e0fefe9-90c8-4cfa-bbe7-610effd92af5 req-a8713f6c-b61c-41e8-a6c5-0a26bc4f7da2 service nova] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Refreshing instance network info cache due to event network-changed-696b0179-1272-4ae7-ab15-727fe4682c42. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1077.200151] env[61970]: DEBUG oslo_concurrency.lockutils [req-6e0fefe9-90c8-4cfa-bbe7-610effd92af5 req-a8713f6c-b61c-41e8-a6c5-0a26bc4f7da2 service nova] Acquiring lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.200151] env[61970]: DEBUG oslo_concurrency.lockutils [req-6e0fefe9-90c8-4cfa-bbe7-610effd92af5 req-a8713f6c-b61c-41e8-a6c5-0a26bc4f7da2 service nova] Acquired lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.200151] env[61970]: DEBUG nova.network.neutron [req-6e0fefe9-90c8-4cfa-bbe7-610effd92af5 req-a8713f6c-b61c-41e8-a6c5-0a26bc4f7da2 service nova] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Refreshing network info cache for port 696b0179-1272-4ae7-ab15-727fe4682c42 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1077.901721] env[61970]: DEBUG nova.network.neutron [req-6e0fefe9-90c8-4cfa-bbe7-610effd92af5 req-a8713f6c-b61c-41e8-a6c5-0a26bc4f7da2 service nova] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updated VIF entry in instance network info cache for port 696b0179-1272-4ae7-ab15-727fe4682c42. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1077.902133] env[61970]: DEBUG nova.network.neutron [req-6e0fefe9-90c8-4cfa-bbe7-610effd92af5 req-a8713f6c-b61c-41e8-a6c5-0a26bc4f7da2 service nova] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating instance_info_cache with network_info: [{"id": "696b0179-1272-4ae7-ab15-727fe4682c42", "address": "fa:16:3e:9e:3f:74", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap696b0179-12", "ovs_interfaceid": "696b0179-1272-4ae7-ab15-727fe4682c42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.015636] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6830c8-69a2-4836-973a-44de9b018ada {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.023353] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2f53ac-a9c9-473d-a432-eabcff9eb1f8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.055269] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79302894-2098-4a51-bd78-af6bd11f73cb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.062291] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af61c6ea-87d6-4a75-ae61-41a8669663a7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.074914] env[61970]: DEBUG nova.compute.provider_tree [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.404908] env[61970]: DEBUG oslo_concurrency.lockutils [req-6e0fefe9-90c8-4cfa-bbe7-610effd92af5 req-a8713f6c-b61c-41e8-a6c5-0a26bc4f7da2 service nova] Releasing lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.579495] env[61970]: DEBUG nova.scheduler.client.report [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1079.087758] env[61970]: DEBUG oslo_concurrency.lockutils [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.170s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.088414] env[61970]: DEBUG nova.compute.manager [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1079.593778] env[61970]: DEBUG nova.compute.utils [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1079.595777] env[61970]: DEBUG nova.compute.manager [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1079.595958] env[61970]: DEBUG nova.network.neutron [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1079.635925] env[61970]: DEBUG nova.policy [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '67a4e38db9894750b16edcf7f188220f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '19dae2316adb4dc980bd972c8c40c034', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 1079.932292] env[61970]: DEBUG nova.network.neutron [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Successfully created port: 8aabb66f-904e-4c84-b459-c3c3ac76d349 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1080.097349] env[61970]: DEBUG nova.compute.manager [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1081.107720] env[61970]: DEBUG nova.compute.manager [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1081.132528] env[61970]: DEBUG nova.virt.hardware [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1081.132792] env[61970]: DEBUG nova.virt.hardware [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1081.132955] env[61970]: DEBUG nova.virt.hardware [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1081.133168] env[61970]: DEBUG nova.virt.hardware [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1081.133319] env[61970]: DEBUG nova.virt.hardware [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1081.133472] env[61970]: DEBUG nova.virt.hardware [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1081.133687] env[61970]: DEBUG nova.virt.hardware [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1081.133852] env[61970]: DEBUG nova.virt.hardware [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1081.134037] env[61970]: DEBUG nova.virt.hardware [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1081.134212] env[61970]: DEBUG nova.virt.hardware [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1081.134423] env[61970]: DEBUG nova.virt.hardware [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1081.135326] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2765ad44-8d13-4c51-870a-87dbb91676d0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.145179] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a98c0a-337d-4bbe-9439-4094b906b5c6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.306476] env[61970]: DEBUG nova.compute.manager [req-407cca27-910c-4af9-b7ee-26c5320ca285 req-47d5a346-0e0a-43af-a298-2007443ead6a service nova] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Received event network-vif-plugged-8aabb66f-904e-4c84-b459-c3c3ac76d349 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1081.306708] env[61970]: DEBUG oslo_concurrency.lockutils [req-407cca27-910c-4af9-b7ee-26c5320ca285 req-47d5a346-0e0a-43af-a298-2007443ead6a service nova] Acquiring lock "df0a2338-07e0-4148-8729-a290b88207b3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.306928] env[61970]: DEBUG oslo_concurrency.lockutils [req-407cca27-910c-4af9-b7ee-26c5320ca285 req-47d5a346-0e0a-43af-a298-2007443ead6a service nova] Lock "df0a2338-07e0-4148-8729-a290b88207b3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.307185] env[61970]: DEBUG oslo_concurrency.lockutils [req-407cca27-910c-4af9-b7ee-26c5320ca285 req-47d5a346-0e0a-43af-a298-2007443ead6a service nova] Lock "df0a2338-07e0-4148-8729-a290b88207b3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.307292] env[61970]: DEBUG nova.compute.manager [req-407cca27-910c-4af9-b7ee-26c5320ca285 req-47d5a346-0e0a-43af-a298-2007443ead6a service nova] [instance: df0a2338-07e0-4148-8729-a290b88207b3] No waiting events found dispatching network-vif-plugged-8aabb66f-904e-4c84-b459-c3c3ac76d349 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1081.307449] env[61970]: WARNING nova.compute.manager [req-407cca27-910c-4af9-b7ee-26c5320ca285 req-47d5a346-0e0a-43af-a298-2007443ead6a service nova] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Received unexpected event network-vif-plugged-8aabb66f-904e-4c84-b459-c3c3ac76d349 for instance with vm_state building and task_state spawning. [ 1081.388050] env[61970]: DEBUG nova.network.neutron [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Successfully updated port: 8aabb66f-904e-4c84-b459-c3c3ac76d349 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1081.890463] env[61970]: DEBUG oslo_concurrency.lockutils [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "refresh_cache-df0a2338-07e0-4148-8729-a290b88207b3" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.890631] env[61970]: DEBUG oslo_concurrency.lockutils [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquired lock "refresh_cache-df0a2338-07e0-4148-8729-a290b88207b3" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.890703] env[61970]: DEBUG nova.network.neutron [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1082.424091] env[61970]: DEBUG nova.network.neutron [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1082.540659] env[61970]: DEBUG nova.network.neutron [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Updating instance_info_cache with network_info: [{"id": "8aabb66f-904e-4c84-b459-c3c3ac76d349", "address": "fa:16:3e:db:48:e5", "network": {"id": "42def9ae-4eed-41e5-a538-81eaced2cdd4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2047751481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19dae2316adb4dc980bd972c8c40c034", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c1b8689-a9b4-4972-beb9-6a1c8de1dc88", "external-id": "nsx-vlan-transportzone-455", "segmentation_id": 455, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8aabb66f-90", "ovs_interfaceid": "8aabb66f-904e-4c84-b459-c3c3ac76d349", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.043380] env[61970]: DEBUG oslo_concurrency.lockutils [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Releasing lock "refresh_cache-df0a2338-07e0-4148-8729-a290b88207b3" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.043751] env[61970]: DEBUG nova.compute.manager [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Instance network_info: |[{"id": "8aabb66f-904e-4c84-b459-c3c3ac76d349", "address": "fa:16:3e:db:48:e5", "network": {"id": "42def9ae-4eed-41e5-a538-81eaced2cdd4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2047751481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19dae2316adb4dc980bd972c8c40c034", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c1b8689-a9b4-4972-beb9-6a1c8de1dc88", "external-id": "nsx-vlan-transportzone-455", "segmentation_id": 455, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8aabb66f-90", "ovs_interfaceid": "8aabb66f-904e-4c84-b459-c3c3ac76d349", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1083.044277] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:48:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9c1b8689-a9b4-4972-beb9-6a1c8de1dc88', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8aabb66f-904e-4c84-b459-c3c3ac76d349', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1083.051694] env[61970]: DEBUG oslo.service.loopingcall [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1083.051958] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1083.052251] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-314ca780-caba-443c-8c78-83bf7ed0eed4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.072719] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1083.072719] env[61970]: value = "task-1356175" [ 1083.072719] env[61970]: _type = "Task" [ 1083.072719] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.080193] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356175, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.333279] env[61970]: DEBUG nova.compute.manager [req-19f0e258-34dc-4097-ae70-9c8838dbcf53 req-85be8973-350a-4c0b-8578-0cdb447fdafd service nova] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Received event network-changed-8aabb66f-904e-4c84-b459-c3c3ac76d349 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1083.333500] env[61970]: DEBUG nova.compute.manager [req-19f0e258-34dc-4097-ae70-9c8838dbcf53 req-85be8973-350a-4c0b-8578-0cdb447fdafd service nova] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Refreshing instance network info cache due to event network-changed-8aabb66f-904e-4c84-b459-c3c3ac76d349. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1083.333617] env[61970]: DEBUG oslo_concurrency.lockutils [req-19f0e258-34dc-4097-ae70-9c8838dbcf53 req-85be8973-350a-4c0b-8578-0cdb447fdafd service nova] Acquiring lock "refresh_cache-df0a2338-07e0-4148-8729-a290b88207b3" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.333750] env[61970]: DEBUG oslo_concurrency.lockutils [req-19f0e258-34dc-4097-ae70-9c8838dbcf53 req-85be8973-350a-4c0b-8578-0cdb447fdafd service nova] Acquired lock "refresh_cache-df0a2338-07e0-4148-8729-a290b88207b3" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.333918] env[61970]: DEBUG nova.network.neutron [req-19f0e258-34dc-4097-ae70-9c8838dbcf53 req-85be8973-350a-4c0b-8578-0cdb447fdafd service nova] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Refreshing network info cache for port 8aabb66f-904e-4c84-b459-c3c3ac76d349 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1083.582633] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356175, 'name': CreateVM_Task, 'duration_secs': 0.298136} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.582969] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1083.583489] env[61970]: DEBUG oslo_concurrency.lockutils [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.583662] env[61970]: DEBUG oslo_concurrency.lockutils [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.583986] env[61970]: DEBUG oslo_concurrency.lockutils [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1083.584258] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0323e817-c624-47df-8708-4fc1d60edafb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.588890] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1083.588890] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ce03c6-2bef-6021-8c84-c20e1f3b293b" [ 1083.588890] env[61970]: _type = "Task" [ 1083.588890] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.597185] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ce03c6-2bef-6021-8c84-c20e1f3b293b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.056271] env[61970]: DEBUG nova.network.neutron [req-19f0e258-34dc-4097-ae70-9c8838dbcf53 req-85be8973-350a-4c0b-8578-0cdb447fdafd service nova] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Updated VIF entry in instance network info cache for port 8aabb66f-904e-4c84-b459-c3c3ac76d349. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1084.056691] env[61970]: DEBUG nova.network.neutron [req-19f0e258-34dc-4097-ae70-9c8838dbcf53 req-85be8973-350a-4c0b-8578-0cdb447fdafd service nova] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Updating instance_info_cache with network_info: [{"id": "8aabb66f-904e-4c84-b459-c3c3ac76d349", "address": "fa:16:3e:db:48:e5", "network": {"id": "42def9ae-4eed-41e5-a538-81eaced2cdd4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2047751481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19dae2316adb4dc980bd972c8c40c034", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c1b8689-a9b4-4972-beb9-6a1c8de1dc88", "external-id": "nsx-vlan-transportzone-455", "segmentation_id": 455, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8aabb66f-90", "ovs_interfaceid": "8aabb66f-904e-4c84-b459-c3c3ac76d349", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.098440] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ce03c6-2bef-6021-8c84-c20e1f3b293b, 'name': SearchDatastore_Task, 'duration_secs': 0.009056} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.099021] env[61970]: DEBUG oslo_concurrency.lockutils [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.099272] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1084.099502] env[61970]: DEBUG oslo_concurrency.lockutils [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.099653] env[61970]: DEBUG oslo_concurrency.lockutils [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.099831] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1084.100098] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6bbaf0db-50e5-4be9-99f1-5e2509a2766c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.109279] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1084.109461] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1084.110185] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb465845-dd4a-44fd-ae99-467a75b7aa76 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.114991] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1084.114991] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5242c63c-98c2-006e-85e9-dc744f6a7ec9" [ 1084.114991] env[61970]: _type = "Task" [ 1084.114991] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.122372] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5242c63c-98c2-006e-85e9-dc744f6a7ec9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.559742] env[61970]: DEBUG oslo_concurrency.lockutils [req-19f0e258-34dc-4097-ae70-9c8838dbcf53 req-85be8973-350a-4c0b-8578-0cdb447fdafd service nova] Releasing lock "refresh_cache-df0a2338-07e0-4148-8729-a290b88207b3" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.626027] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5242c63c-98c2-006e-85e9-dc744f6a7ec9, 'name': SearchDatastore_Task, 'duration_secs': 0.007817} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.626027] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f400f676-1e1f-4388-a4d0-267c5b7db106 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.631293] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1084.631293] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5298b29c-2544-aa15-a235-a3498bb410eb" [ 1084.631293] env[61970]: _type = "Task" [ 1084.631293] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.638800] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5298b29c-2544-aa15-a235-a3498bb410eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.142251] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5298b29c-2544-aa15-a235-a3498bb410eb, 'name': SearchDatastore_Task, 'duration_secs': 0.009102} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.142251] env[61970]: DEBUG oslo_concurrency.lockutils [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.142453] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] df0a2338-07e0-4148-8729-a290b88207b3/df0a2338-07e0-4148-8729-a290b88207b3.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1085.142700] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5e0498ca-d753-464e-a172-f3dd4f7a6f99 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.149065] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1085.149065] env[61970]: value = "task-1356177" [ 1085.149065] env[61970]: _type = "Task" [ 1085.149065] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.156651] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356177, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.659080] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356177, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473455} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.659418] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] df0a2338-07e0-4148-8729-a290b88207b3/df0a2338-07e0-4148-8729-a290b88207b3.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1085.659564] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1085.659815] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d2169ea9-3443-46a7-8579-0e5feaafd07e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.666157] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1085.666157] env[61970]: value = "task-1356178" [ 1085.666157] env[61970]: _type = "Task" [ 1085.666157] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.673348] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356178, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.175597] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356178, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058535} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.175885] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1086.176651] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af641643-8380-453e-b825-0d3695f8ee4f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.198090] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] df0a2338-07e0-4148-8729-a290b88207b3/df0a2338-07e0-4148-8729-a290b88207b3.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1086.198363] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94252a0e-ae61-4851-84db-67ce443b84e0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.219227] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1086.219227] env[61970]: value = "task-1356180" [ 1086.219227] env[61970]: _type = "Task" [ 1086.219227] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.227478] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356180, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.260315] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.260531] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.260708] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.260896] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.261057] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Cleaning up deleted instances {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1086.728985] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356180, 'name': ReconfigVM_Task, 'duration_secs': 0.344824} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.729354] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Reconfigured VM instance instance-00000069 to attach disk [datastore2] df0a2338-07e0-4148-8729-a290b88207b3/df0a2338-07e0-4148-8729-a290b88207b3.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1086.729890] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b744202e-7e96-4bfe-a528-68070ce2af5d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.735296] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1086.735296] env[61970]: value = "task-1356181" [ 1086.735296] env[61970]: _type = "Task" [ 1086.735296] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.743729] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356181, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.772694] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] There are 40 instances to clean {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1086.772957] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: ef73a1ea-ffc7-41a0-ade7-02ecc5e9eb97] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.244682] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356181, 'name': Rename_Task, 'duration_secs': 0.223073} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.244952] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1087.245218] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-56d04472-29d0-4450-b8f6-f9e071a1c4bc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.250642] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1087.250642] env[61970]: value = "task-1356182" [ 1087.250642] env[61970]: _type = "Task" [ 1087.250642] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.262735] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356182, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.276304] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 444939f3-f0d7-4af6-a3fa-40c0a8459839] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.759831] env[61970]: DEBUG oslo_vmware.api [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356182, 'name': PowerOnVM_Task, 'duration_secs': 0.435322} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.760132] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1087.760328] env[61970]: INFO nova.compute.manager [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Took 6.65 seconds to spawn the instance on the hypervisor. [ 1087.760509] env[61970]: DEBUG nova.compute.manager [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1087.761293] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe19ab4-1be6-48d3-8e7c-be2b6b1e789e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.779310] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 19c0bfcb-1964-4719-98b1-d9334751ace2] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.279071] env[61970]: INFO nova.compute.manager [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Took 11.38 seconds to build instance. [ 1088.282413] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 31f9d37b-f4ff-4a8b-9477-8253f8305020] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.780876] env[61970]: DEBUG oslo_concurrency.lockutils [None req-804f12d0-10cd-4ccb-82e8-f710496cdb52 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "df0a2338-07e0-4148-8729-a290b88207b3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.890s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.785475] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: d38bde37-994b-4d58-b4b1-5f79b59fd2aa] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.112502] env[61970]: DEBUG nova.compute.manager [req-bf7999e5-677c-4154-b60a-8546f724f7ae req-0f9e640d-fda7-4bdb-be6b-3ec2d04393d1 service nova] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Received event network-changed-8aabb66f-904e-4c84-b459-c3c3ac76d349 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1089.112732] env[61970]: DEBUG nova.compute.manager [req-bf7999e5-677c-4154-b60a-8546f724f7ae req-0f9e640d-fda7-4bdb-be6b-3ec2d04393d1 service nova] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Refreshing instance network info cache due to event network-changed-8aabb66f-904e-4c84-b459-c3c3ac76d349. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1089.112966] env[61970]: DEBUG oslo_concurrency.lockutils [req-bf7999e5-677c-4154-b60a-8546f724f7ae req-0f9e640d-fda7-4bdb-be6b-3ec2d04393d1 service nova] Acquiring lock "refresh_cache-df0a2338-07e0-4148-8729-a290b88207b3" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.113159] env[61970]: DEBUG oslo_concurrency.lockutils [req-bf7999e5-677c-4154-b60a-8546f724f7ae req-0f9e640d-fda7-4bdb-be6b-3ec2d04393d1 service nova] Acquired lock "refresh_cache-df0a2338-07e0-4148-8729-a290b88207b3" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.113335] env[61970]: DEBUG nova.network.neutron [req-bf7999e5-677c-4154-b60a-8546f724f7ae req-0f9e640d-fda7-4bdb-be6b-3ec2d04393d1 service nova] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Refreshing network info cache for port 8aabb66f-904e-4c84-b459-c3c3ac76d349 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1089.288893] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 00e3287d-24fb-4ea9-b212-54215828d5df] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.792797] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 3b1e8446-7862-4532-b2ed-611b63dbf391] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.013606] env[61970]: DEBUG nova.network.neutron [req-bf7999e5-677c-4154-b60a-8546f724f7ae req-0f9e640d-fda7-4bdb-be6b-3ec2d04393d1 service nova] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Updated VIF entry in instance network info cache for port 8aabb66f-904e-4c84-b459-c3c3ac76d349. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1090.014090] env[61970]: DEBUG nova.network.neutron [req-bf7999e5-677c-4154-b60a-8546f724f7ae req-0f9e640d-fda7-4bdb-be6b-3ec2d04393d1 service nova] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Updating instance_info_cache with network_info: [{"id": "8aabb66f-904e-4c84-b459-c3c3ac76d349", "address": "fa:16:3e:db:48:e5", "network": {"id": "42def9ae-4eed-41e5-a538-81eaced2cdd4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2047751481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19dae2316adb4dc980bd972c8c40c034", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c1b8689-a9b4-4972-beb9-6a1c8de1dc88", "external-id": "nsx-vlan-transportzone-455", "segmentation_id": 455, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8aabb66f-90", "ovs_interfaceid": "8aabb66f-904e-4c84-b459-c3c3ac76d349", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.296032] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: bc948a41-1a22-46fa-a76a-50b4c4afa4b7] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.517363] env[61970]: DEBUG oslo_concurrency.lockutils [req-bf7999e5-677c-4154-b60a-8546f724f7ae req-0f9e640d-fda7-4bdb-be6b-3ec2d04393d1 service nova] Releasing lock "refresh_cache-df0a2338-07e0-4148-8729-a290b88207b3" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.800218] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 4d093692-07d3-49aa-8b0a-fb52bd9c0ffc] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.303964] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 479342e0-7627-4dda-a589-a7bc5a5c4270] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.807977] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: c4f8a3f1-4e30-42b0-8c18-8fabc7957ca8] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.312406] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: b9be8292-378b-4187-85dc-8d01e817faa3] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.816302] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 0d21db81-a53f-4edf-a3a6-f1949cba2499] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.320915] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: c4ec1797-d939-4c24-9314-ce9c74eb5aa1] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.824196] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: fa8bf4b0-6955-4932-a85d-5a33947ae594] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.326867] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: ba4df05e-7fdd-43b4-9c4f-d52cf5319cdc] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.830619] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: dc38f9da-9f82-4707-9764-b28aa83a8ece] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.334418] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: e4e85cff-49e7-4306-b5bc-15a798fc7e7e] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.664521] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.664781] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.837769] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 3b346c68-53a9-4ceb-83a3-9e4bce9610b1] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.166963] env[61970]: DEBUG nova.compute.manager [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1096.341404] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: ebb92a5b-f635-4504-a57f-395de514015a] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.767468] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.767737] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.769389] env[61970]: INFO nova.compute.claims [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1096.844731] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 7f59000a-94d9-45b6-aa7a-300d95793615] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.350044] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: d0ad9689-bdb7-4d68-b19a-3f805ab4612e] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.852846] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 8ce9b4e2-b72e-43d4-8a10-3c2f20d5ab20] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.859380] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-598a1040-bdca-4b2c-8d44-54f797203ae6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.867848] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2268701e-ea39-4378-89a1-e2bfb141fa2a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.898688] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e7baf3-13b9-4a3f-958c-dda8808b0fa4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.906225] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3399bc2-5132-405d-9f07-a5596cf8ea08 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.920798] env[61970]: DEBUG nova.compute.provider_tree [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1098.355676] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 75762ba3-f129-424e-a7cb-962785dfab2a] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.423861] env[61970]: DEBUG nova.scheduler.client.report [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1098.859023] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: e2185ed0-2bc1-4718-b47d-57150b5e60ba] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.928047] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.160s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.928482] env[61970]: DEBUG nova.compute.manager [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1099.362138] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 99200b46-7eb7-4ca2-b352-b91b4ad076af] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.434074] env[61970]: DEBUG nova.compute.utils [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1099.434921] env[61970]: DEBUG nova.compute.manager [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1099.435106] env[61970]: DEBUG nova.network.neutron [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1099.481409] env[61970]: DEBUG nova.policy [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '223ba28a2a204f8ba9e6d2e0f876ddd2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '28a078b8af9c469eb279be4da7459166', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 1099.801890] env[61970]: DEBUG nova.network.neutron [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Successfully created port: 8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1099.865426] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 15ad6d33-0ab2-4598-bfa7-0f70bca5f8fa] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.940022] env[61970]: DEBUG nova.compute.manager [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1100.369378] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 2d280f75-ed84-4db3-9067-8ccc719b0d2e] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1100.443457] env[61970]: INFO nova.virt.block_device [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Booting with volume ed3d65b8-03da-4484-8a98-b0e8211e6d54 at /dev/sda [ 1100.479350] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-214a0dd6-99d9-44c3-941f-bfb7269b0de4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.488433] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4a478e-23d2-452c-bc37-aeb3ee911a92 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.516613] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6ef8eedf-642c-4b83-bee2-1525e47aae8a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.525011] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b1bfc2e-1b75-4951-b779-09536e591695 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.554452] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d265f9-4219-4f4b-80cf-910eb31ef272 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.561141] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe97a72-d32d-4c6c-8855-6a28f6923d1c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.575040] env[61970]: DEBUG nova.virt.block_device [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updating existing volume attachment record: 5c933750-2edf-4c39-bd1b-3d49ad452b40 {{(pid=61970) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1100.872734] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 035fb797-21c1-4af9-8e66-deee8e9d083a] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1100.893661] env[61970]: DEBUG oslo_concurrency.lockutils [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.893899] env[61970]: DEBUG oslo_concurrency.lockutils [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.159811] env[61970]: DEBUG nova.compute.manager [req-2bd25906-5330-49e6-a126-7e1641daa83b req-3206cf6a-c764-4298-9fc3-eb08f5435ca9 service nova] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Received event network-vif-plugged-8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1101.160178] env[61970]: DEBUG oslo_concurrency.lockutils [req-2bd25906-5330-49e6-a126-7e1641daa83b req-3206cf6a-c764-4298-9fc3-eb08f5435ca9 service nova] Acquiring lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.160500] env[61970]: DEBUG oslo_concurrency.lockutils [req-2bd25906-5330-49e6-a126-7e1641daa83b req-3206cf6a-c764-4298-9fc3-eb08f5435ca9 service nova] Lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.160694] env[61970]: DEBUG oslo_concurrency.lockutils [req-2bd25906-5330-49e6-a126-7e1641daa83b req-3206cf6a-c764-4298-9fc3-eb08f5435ca9 service nova] Lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.160874] env[61970]: DEBUG nova.compute.manager [req-2bd25906-5330-49e6-a126-7e1641daa83b req-3206cf6a-c764-4298-9fc3-eb08f5435ca9 service nova] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] No waiting events found dispatching network-vif-plugged-8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1101.161061] env[61970]: WARNING nova.compute.manager [req-2bd25906-5330-49e6-a126-7e1641daa83b req-3206cf6a-c764-4298-9fc3-eb08f5435ca9 service nova] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Received unexpected event network-vif-plugged-8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d for instance with vm_state building and task_state block_device_mapping. [ 1101.377718] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: e1415314-f137-4ee3-a065-3d875ef9e7ff] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.397046] env[61970]: DEBUG nova.compute.utils [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1101.731321] env[61970]: DEBUG nova.network.neutron [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Successfully updated port: 8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1101.752434] env[61970]: DEBUG nova.compute.manager [req-a5f6bc81-8495-4538-8fd8-32cc3a9ca106 req-0aabc795-ab79-4c95-bf27-cbae0d9f5409 service nova] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Received event network-changed-8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1101.752653] env[61970]: DEBUG nova.compute.manager [req-a5f6bc81-8495-4538-8fd8-32cc3a9ca106 req-0aabc795-ab79-4c95-bf27-cbae0d9f5409 service nova] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Refreshing instance network info cache due to event network-changed-8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1101.752969] env[61970]: DEBUG oslo_concurrency.lockutils [req-a5f6bc81-8495-4538-8fd8-32cc3a9ca106 req-0aabc795-ab79-4c95-bf27-cbae0d9f5409 service nova] Acquiring lock "refresh_cache-a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.753115] env[61970]: DEBUG oslo_concurrency.lockutils [req-a5f6bc81-8495-4538-8fd8-32cc3a9ca106 req-0aabc795-ab79-4c95-bf27-cbae0d9f5409 service nova] Acquired lock "refresh_cache-a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.753282] env[61970]: DEBUG nova.network.neutron [req-a5f6bc81-8495-4538-8fd8-32cc3a9ca106 req-0aabc795-ab79-4c95-bf27-cbae0d9f5409 service nova] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Refreshing network info cache for port 8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1101.881301] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: d09dbde2-5daa-4ae9-9fd5-bfbb03fba136] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.899513] env[61970]: DEBUG oslo_concurrency.lockutils [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.237474] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "refresh_cache-a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1102.283458] env[61970]: DEBUG nova.network.neutron [req-a5f6bc81-8495-4538-8fd8-32cc3a9ca106 req-0aabc795-ab79-4c95-bf27-cbae0d9f5409 service nova] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1102.355483] env[61970]: DEBUG nova.network.neutron [req-a5f6bc81-8495-4538-8fd8-32cc3a9ca106 req-0aabc795-ab79-4c95-bf27-cbae0d9f5409 service nova] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.384619] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 6838c6a4-8241-4b6f-ad58-3acfd7afdbf4] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1102.656878] env[61970]: DEBUG nova.compute.manager [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1102.657440] env[61970]: DEBUG nova.virt.hardware [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1102.657698] env[61970]: DEBUG nova.virt.hardware [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1102.657882] env[61970]: DEBUG nova.virt.hardware [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1102.658092] env[61970]: DEBUG nova.virt.hardware [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1102.658248] env[61970]: DEBUG nova.virt.hardware [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1102.658408] env[61970]: DEBUG nova.virt.hardware [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1102.658615] env[61970]: DEBUG nova.virt.hardware [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1102.658778] env[61970]: DEBUG nova.virt.hardware [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1102.658950] env[61970]: DEBUG nova.virt.hardware [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1102.659224] env[61970]: DEBUG nova.virt.hardware [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1102.659428] env[61970]: DEBUG nova.virt.hardware [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1102.660314] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2386955f-6105-4ee1-99f1-ae7a24e975dc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.669217] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f96377c-2c93-498e-adf7-60f5acfa378f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.858694] env[61970]: DEBUG oslo_concurrency.lockutils [req-a5f6bc81-8495-4538-8fd8-32cc3a9ca106 req-0aabc795-ab79-4c95-bf27-cbae0d9f5409 service nova] Releasing lock "refresh_cache-a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1102.859044] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "refresh_cache-a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.859215] env[61970]: DEBUG nova.network.neutron [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1102.887193] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: b385407b-1bdd-4c53-907c-cb4c8ce16cc7] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1102.962788] env[61970]: DEBUG oslo_concurrency.lockutils [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.963064] env[61970]: DEBUG oslo_concurrency.lockutils [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.963344] env[61970]: INFO nova.compute.manager [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Attaching volume 3ff6c6bf-861a-44c7-8436-f62c37e40dfb to /dev/sdb [ 1102.992224] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46bb8ed8-3a1b-43d6-a96e-7baac7bc85e0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.999734] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d69596-95b5-44a3-bb3d-2aff73df9f2b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.012760] env[61970]: DEBUG nova.virt.block_device [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Updating existing volume attachment record: 149f2c02-2bca-4480-8a29-255717c4e84a {{(pid=61970) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1103.390810] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: b1cc1cc2-15d6-459d-9529-e592ddb225ac] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1103.393996] env[61970]: DEBUG nova.network.neutron [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1103.531630] env[61970]: DEBUG nova.network.neutron [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updating instance_info_cache with network_info: [{"id": "8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d", "address": "fa:16:3e:48:95:bf", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d2476c8-b7", "ovs_interfaceid": "8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.897009] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: b19cb516-b163-4bed-ba5b-139a0a18fc05] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1104.034331] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "refresh_cache-a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.034671] env[61970]: DEBUG nova.compute.manager [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Instance network_info: |[{"id": "8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d", "address": "fa:16:3e:48:95:bf", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d2476c8-b7", "ovs_interfaceid": "8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1104.035246] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:95:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b356db78-99c7-4464-822c-fc7e193f7878', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1104.043091] env[61970]: DEBUG oslo.service.loopingcall [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1104.043341] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1104.043619] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e966e792-a7e8-4112-b408-da58a9a99596 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.065078] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1104.065078] env[61970]: value = "task-1356187" [ 1104.065078] env[61970]: _type = "Task" [ 1104.065078] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.072398] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356187, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.400097] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: e2f51299-56fe-46cc-9e08-2246178b9db7] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1104.575417] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356187, 'name': CreateVM_Task, 'duration_secs': 0.340881} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.575561] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1104.576271] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': True, 'attachment_id': '5c933750-2edf-4c39-bd1b-3d49ad452b40', 'boot_index': 0, 'device_type': None, 'mount_device': '/dev/sda', 'disk_bus': None, 'guest_format': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288883', 'volume_id': 'ed3d65b8-03da-4484-8a98-b0e8211e6d54', 'name': 'volume-ed3d65b8-03da-4484-8a98-b0e8211e6d54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8', 'attached_at': '', 'detached_at': '', 'volume_id': 'ed3d65b8-03da-4484-8a98-b0e8211e6d54', 'serial': 'ed3d65b8-03da-4484-8a98-b0e8211e6d54'}, 'volume_type': None}], 'swap': None} {{(pid=61970) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1104.576490] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Root volume attach. Driver type: vmdk {{(pid=61970) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1104.577277] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc4599d7-4516-4c4d-b26e-d86123e116e1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.584990] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3848fd87-68ca-4722-ac49-27a0dcf707e8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.590887] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2025a024-7e15-41c1-a214-85da8c27b4c4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.596316] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-14764a51-a2cb-4eaa-879d-289bc4c45c5a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.602735] env[61970]: DEBUG oslo_vmware.api [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1104.602735] env[61970]: value = "task-1356188" [ 1104.602735] env[61970]: _type = "Task" [ 1104.602735] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.609913] env[61970]: DEBUG oslo_vmware.api [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356188, 'name': RelocateVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.903954] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 6345d239-fbae-4f3a-9f61-79e10539ff0b] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1105.113219] env[61970]: DEBUG oslo_vmware.api [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356188, 'name': RelocateVM_Task, 'duration_secs': 0.380996} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.113494] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Volume attach. Driver type: vmdk {{(pid=61970) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1105.113664] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288883', 'volume_id': 'ed3d65b8-03da-4484-8a98-b0e8211e6d54', 'name': 'volume-ed3d65b8-03da-4484-8a98-b0e8211e6d54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8', 'attached_at': '', 'detached_at': '', 'volume_id': 'ed3d65b8-03da-4484-8a98-b0e8211e6d54', 'serial': 'ed3d65b8-03da-4484-8a98-b0e8211e6d54'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1105.114489] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d8f28d-da77-47aa-ae13-48e6dd97b4d6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.129323] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b6e3de-db04-49e5-b365-531db5e3027a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.150195] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] volume-ed3d65b8-03da-4484-8a98-b0e8211e6d54/volume-ed3d65b8-03da-4484-8a98-b0e8211e6d54.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1105.150473] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e48f2dda-cdd6-4123-8b74-0e7597aae898 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.169029] env[61970]: DEBUG oslo_vmware.api [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1105.169029] env[61970]: value = "task-1356189" [ 1105.169029] env[61970]: _type = "Task" [ 1105.169029] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.176037] env[61970]: DEBUG oslo_vmware.api [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356189, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.407378] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 5d4d97e1-5812-4884-a1ea-6eb0cee591c0] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1105.678761] env[61970]: DEBUG oslo_vmware.api [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356189, 'name': ReconfigVM_Task, 'duration_secs': 0.32314} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.679010] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Reconfigured VM instance instance-0000006a to attach disk [datastore1] volume-ed3d65b8-03da-4484-8a98-b0e8211e6d54/volume-ed3d65b8-03da-4484-8a98-b0e8211e6d54.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1105.683661] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5bfe6011-2051-4927-b39b-47b588e8b60b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.698748] env[61970]: DEBUG oslo_vmware.api [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1105.698748] env[61970]: value = "task-1356191" [ 1105.698748] env[61970]: _type = "Task" [ 1105.698748] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.707400] env[61970]: DEBUG oslo_vmware.api [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356191, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.911579] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: a78b63a5-6bb8-4271-90d8-1e86fb29db4f] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.208922] env[61970]: DEBUG oslo_vmware.api [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356191, 'name': ReconfigVM_Task, 'duration_secs': 0.182662} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.209193] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288883', 'volume_id': 'ed3d65b8-03da-4484-8a98-b0e8211e6d54', 'name': 'volume-ed3d65b8-03da-4484-8a98-b0e8211e6d54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8', 'attached_at': '', 'detached_at': '', 'volume_id': 'ed3d65b8-03da-4484-8a98-b0e8211e6d54', 'serial': 'ed3d65b8-03da-4484-8a98-b0e8211e6d54'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1106.209708] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-95cdb007-5d9a-4fb7-b885-64a445f40d98 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.215415] env[61970]: DEBUG oslo_vmware.api [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1106.215415] env[61970]: value = "task-1356192" [ 1106.215415] env[61970]: _type = "Task" [ 1106.215415] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.223066] env[61970]: DEBUG oslo_vmware.api [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356192, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.414960] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 7d82cf87-1019-4ba6-83f4-fcd2d3c2aec0] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.725407] env[61970]: DEBUG oslo_vmware.api [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356192, 'name': Rename_Task, 'duration_secs': 0.135455} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.725742] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1106.726040] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07e4e555-c3cc-439c-99ac-791d9cb807c1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.731950] env[61970]: DEBUG oslo_vmware.api [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1106.731950] env[61970]: value = "task-1356193" [ 1106.731950] env[61970]: _type = "Task" [ 1106.731950] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.738742] env[61970]: DEBUG oslo_vmware.api [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356193, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.918760] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.919156] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Cleaning up deleted instances with incomplete migration {{(pid=61970) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1107.241935] env[61970]: DEBUG oslo_vmware.api [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356193, 'name': PowerOnVM_Task, 'duration_secs': 0.465214} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.242231] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1107.242440] env[61970]: INFO nova.compute.manager [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Took 4.59 seconds to spawn the instance on the hypervisor. [ 1107.242622] env[61970]: DEBUG nova.compute.manager [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1107.243389] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24f01df-5991-448c-8f25-9421ed07ba34 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.421894] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.554764] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Volume attach. Driver type: vmdk {{(pid=61970) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1107.555013] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288884', 'volume_id': '3ff6c6bf-861a-44c7-8436-f62c37e40dfb', 'name': 'volume-3ff6c6bf-861a-44c7-8436-f62c37e40dfb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd96208cf-fcc5-4014-ac23-1d56ea97050a', 'attached_at': '', 'detached_at': '', 'volume_id': '3ff6c6bf-861a-44c7-8436-f62c37e40dfb', 'serial': '3ff6c6bf-861a-44c7-8436-f62c37e40dfb'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1107.555909] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deed45c4-4390-40a0-8f04-04d00d17dceb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.572019] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d36413c-32a4-46f3-84c6-1cd0b3783847 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.595180] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] volume-3ff6c6bf-861a-44c7-8436-f62c37e40dfb/volume-3ff6c6bf-861a-44c7-8436-f62c37e40dfb.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1107.595428] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f00bf2e0-4122-4c16-a92b-c84839cac6ac {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.612325] env[61970]: DEBUG oslo_vmware.api [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1107.612325] env[61970]: value = "task-1356194" [ 1107.612325] env[61970]: _type = "Task" [ 1107.612325] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.622197] env[61970]: DEBUG oslo_vmware.api [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356194, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.761472] env[61970]: INFO nova.compute.manager [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Took 11.09 seconds to build instance. [ 1108.123453] env[61970]: DEBUG oslo_vmware.api [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356194, 'name': ReconfigVM_Task, 'duration_secs': 0.337762} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.123771] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Reconfigured VM instance instance-00000064 to attach disk [datastore1] volume-3ff6c6bf-861a-44c7-8436-f62c37e40dfb/volume-3ff6c6bf-861a-44c7-8436-f62c37e40dfb.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1108.128382] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb3e39e6-9585-4a70-85bf-5e4b02c88c9d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.143780] env[61970]: DEBUG oslo_vmware.api [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1108.143780] env[61970]: value = "task-1356195" [ 1108.143780] env[61970]: _type = "Task" [ 1108.143780] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.152078] env[61970]: DEBUG oslo_vmware.api [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356195, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.262909] env[61970]: DEBUG oslo_concurrency.lockutils [None req-de2d5b16-7e50-4be7-ab29-aac363303fba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.598s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.656404] env[61970]: DEBUG oslo_vmware.api [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356195, 'name': ReconfigVM_Task, 'duration_secs': 0.153896} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.656738] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288884', 'volume_id': '3ff6c6bf-861a-44c7-8436-f62c37e40dfb', 'name': 'volume-3ff6c6bf-861a-44c7-8436-f62c37e40dfb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd96208cf-fcc5-4014-ac23-1d56ea97050a', 'attached_at': '', 'detached_at': '', 'volume_id': '3ff6c6bf-861a-44c7-8436-f62c37e40dfb', 'serial': '3ff6c6bf-861a-44c7-8436-f62c37e40dfb'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1108.829479] env[61970]: DEBUG nova.compute.manager [req-bc99b339-aa89-4485-a643-2f86ccfbcda4 req-470c9b03-0687-4f8d-9171-e450d62ba89c service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Received event network-changed-2545e2b0-a432-44ae-a287-0b68ceab7031 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1108.829701] env[61970]: DEBUG nova.compute.manager [req-bc99b339-aa89-4485-a643-2f86ccfbcda4 req-470c9b03-0687-4f8d-9171-e450d62ba89c service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Refreshing instance network info cache due to event network-changed-2545e2b0-a432-44ae-a287-0b68ceab7031. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1108.829921] env[61970]: DEBUG oslo_concurrency.lockutils [req-bc99b339-aa89-4485-a643-2f86ccfbcda4 req-470c9b03-0687-4f8d-9171-e450d62ba89c service nova] Acquiring lock "refresh_cache-0304640a-6c46-4f57-a8b8-5058da2e3de4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.830081] env[61970]: DEBUG oslo_concurrency.lockutils [req-bc99b339-aa89-4485-a643-2f86ccfbcda4 req-470c9b03-0687-4f8d-9171-e450d62ba89c service nova] Acquired lock "refresh_cache-0304640a-6c46-4f57-a8b8-5058da2e3de4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.830295] env[61970]: DEBUG nova.network.neutron [req-bc99b339-aa89-4485-a643-2f86ccfbcda4 req-470c9b03-0687-4f8d-9171-e450d62ba89c service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Refreshing network info cache for port 2545e2b0-a432-44ae-a287-0b68ceab7031 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1108.923386] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1109.429712] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1109.430037] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Starting heal instance info cache {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1109.533099] env[61970]: DEBUG nova.network.neutron [req-bc99b339-aa89-4485-a643-2f86ccfbcda4 req-470c9b03-0687-4f8d-9171-e450d62ba89c service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Updated VIF entry in instance network info cache for port 2545e2b0-a432-44ae-a287-0b68ceab7031. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1109.533470] env[61970]: DEBUG nova.network.neutron [req-bc99b339-aa89-4485-a643-2f86ccfbcda4 req-470c9b03-0687-4f8d-9171-e450d62ba89c service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Updating instance_info_cache with network_info: [{"id": "2545e2b0-a432-44ae-a287-0b68ceab7031", "address": "fa:16:3e:a6:77:ea", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2545e2b0-a4", "ovs_interfaceid": "2545e2b0-a432-44ae-a287-0b68ceab7031", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.696177] env[61970]: DEBUG nova.objects.instance [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lazy-loading 'flavor' on Instance uuid d96208cf-fcc5-4014-ac23-1d56ea97050a {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1110.037069] env[61970]: DEBUG oslo_concurrency.lockutils [req-bc99b339-aa89-4485-a643-2f86ccfbcda4 req-470c9b03-0687-4f8d-9171-e450d62ba89c service nova] Releasing lock "refresh_cache-0304640a-6c46-4f57-a8b8-5058da2e3de4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.202931] env[61970]: DEBUG oslo_concurrency.lockutils [None req-385b629b-d374-43fa-8f2a-81e67f6c25fc tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.239s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.238576] env[61970]: DEBUG nova.compute.manager [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Stashing vm_state: active {{(pid=61970) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1110.756989] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.757261] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.858405] env[61970]: DEBUG nova.compute.manager [req-00ee3c79-707e-46ef-971f-d2c2bdf2a69a req-8719f5b1-74f4-4bd5-96ff-3a6f6610aa08 service nova] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Received event network-changed-8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1110.858654] env[61970]: DEBUG nova.compute.manager [req-00ee3c79-707e-46ef-971f-d2c2bdf2a69a req-8719f5b1-74f4-4bd5-96ff-3a6f6610aa08 service nova] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Refreshing instance network info cache due to event network-changed-8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1110.858876] env[61970]: DEBUG oslo_concurrency.lockutils [req-00ee3c79-707e-46ef-971f-d2c2bdf2a69a req-8719f5b1-74f4-4bd5-96ff-3a6f6610aa08 service nova] Acquiring lock "refresh_cache-a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.859037] env[61970]: DEBUG oslo_concurrency.lockutils [req-00ee3c79-707e-46ef-971f-d2c2bdf2a69a req-8719f5b1-74f4-4bd5-96ff-3a6f6610aa08 service nova] Acquired lock "refresh_cache-a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.859210] env[61970]: DEBUG nova.network.neutron [req-00ee3c79-707e-46ef-971f-d2c2bdf2a69a req-8719f5b1-74f4-4bd5-96ff-3a6f6610aa08 service nova] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Refreshing network info cache for port 8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1111.060840] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.061091] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.262569] env[61970]: INFO nova.compute.claims [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1111.563997] env[61970]: DEBUG nova.compute.utils [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1111.586508] env[61970]: DEBUG nova.network.neutron [req-00ee3c79-707e-46ef-971f-d2c2bdf2a69a req-8719f5b1-74f4-4bd5-96ff-3a6f6610aa08 service nova] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updated VIF entry in instance network info cache for port 8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1111.586508] env[61970]: DEBUG nova.network.neutron [req-00ee3c79-707e-46ef-971f-d2c2bdf2a69a req-8719f5b1-74f4-4bd5-96ff-3a6f6610aa08 service nova] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updating instance_info_cache with network_info: [{"id": "8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d", "address": "fa:16:3e:48:95:bf", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d2476c8-b7", "ovs_interfaceid": "8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.767914] env[61970]: INFO nova.compute.resource_tracker [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updating resource usage from migration a273a020-6905-4994-806d-f9aa5f5c9d19 [ 1111.861081] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b175c7-7829-4ce3-ac72-264b2adfaefe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.868704] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c856ee86-2d3a-4ce6-9c53-9bc3a0c8f4fb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.897342] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3687c01-46e9-4555-ad77-9acec0447c4e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.903880] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634c896a-6ec2-4a64-ab86-74e31f10f230 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.916174] env[61970]: DEBUG nova.compute.provider_tree [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1111.998026] env[61970]: DEBUG oslo_concurrency.lockutils [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.998258] env[61970]: DEBUG oslo_concurrency.lockutils [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.998446] env[61970]: INFO nova.compute.manager [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Shelving [ 1112.066767] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.087040] env[61970]: DEBUG oslo_concurrency.lockutils [req-00ee3c79-707e-46ef-971f-d2c2bdf2a69a req-8719f5b1-74f4-4bd5-96ff-3a6f6610aa08 service nova] Releasing lock "refresh_cache-a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.418935] env[61970]: DEBUG nova.scheduler.client.report [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1112.447861] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Didn't find any instances for network info cache update. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1112.448095] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1112.448274] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1112.448427] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1112.448572] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1112.448701] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61970) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1112.448842] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1112.505361] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1112.505608] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88094c7a-7bf2-46db-bc77-e52df2febb60 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.513514] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1112.513514] env[61970]: value = "task-1356196" [ 1112.513514] env[61970]: _type = "Task" [ 1112.513514] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.521637] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356196, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.834928] env[61970]: DEBUG oslo_concurrency.lockutils [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "5a8cdb36-1845-4806-a6bc-22afb47886e2" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.834928] env[61970]: DEBUG oslo_concurrency.lockutils [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "5a8cdb36-1845-4806-a6bc-22afb47886e2" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.923682] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.166s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.923927] env[61970]: INFO nova.compute.manager [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Migrating [ 1112.952359] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.952505] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.952663] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.952808] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61970) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1112.953891] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88335e87-2b14-4b3f-851c-6e4c87a55f2e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.962176] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad27c286-26a0-4d0e-a51a-d7c171322c0b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.978040] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0cce345-81bd-49e8-8b4f-59e7a6123903 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.984141] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748d9bb6-946c-4a47-9260-7375e279da30 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.013344] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180479MB free_disk=150GB free_vcpus=48 pci_devices=None {{(pid=61970) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1113.013484] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.013669] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.022596] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356196, 'name': PowerOffVM_Task, 'duration_secs': 0.188132} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.022852] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1113.023583] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e73c0b-7eab-4a0f-a820-7628da4c9397 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.042649] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e00a5bed-2cf6-46cc-939d-607b5fd8cdcf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.127155] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.127405] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.127643] env[61970]: INFO nova.compute.manager [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Attaching volume e4b113f3-7ebc-465c-95d5-8d14ceaca257 to /dev/sdc [ 1113.158107] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-957e5baf-cb8e-4100-8d73-7b2693b89768 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.164990] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-116dde30-8996-462d-97a6-4bdc51930010 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.177577] env[61970]: DEBUG nova.virt.block_device [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Updating existing volume attachment record: bd11d65f-54f7-4ecc-8f0f-441c6b350bff {{(pid=61970) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1113.337910] env[61970]: DEBUG nova.compute.utils [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1113.437902] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "refresh_cache-a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.438146] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "refresh_cache-a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.438343] env[61970]: DEBUG nova.network.neutron [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1113.553116] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Creating Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1113.553413] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0ff27702-9cb7-4539-bbf8-af4527a03dee {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.561086] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1113.561086] env[61970]: value = "task-1356198" [ 1113.561086] env[61970]: _type = "Task" [ 1113.561086] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.569828] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356198, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.841424] env[61970]: DEBUG oslo_concurrency.lockutils [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "5a8cdb36-1845-4806-a6bc-22afb47886e2" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.024681] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Applying migration context for instance a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8 as it has an incoming, in-progress migration a273a020-6905-4994-806d-f9aa5f5c9d19. Migration status is pre-migrating {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1114.026215] env[61970]: INFO nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updating resource usage from migration a273a020-6905-4994-806d-f9aa5f5c9d19 [ 1114.048969] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 0304640a-6c46-4f57-a8b8-5058da2e3de4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1114.049145] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance d96208cf-fcc5-4014-ac23-1d56ea97050a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1114.049274] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1114.049393] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 183133c6-7e4e-4dd4-b8e0-cb4f8338264c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1114.049509] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 5a8cdb36-1845-4806-a6bc-22afb47886e2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1114.049622] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance df0a2338-07e0-4148-8729-a290b88207b3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1114.049734] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Migration a273a020-6905-4994-806d-f9aa5f5c9d19 is active on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1114.049846] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1114.050053] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1114.050247] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2112MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1114.072753] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356198, 'name': CreateSnapshot_Task, 'duration_secs': 0.433777} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.073050] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Created Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1114.073849] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f24c66b8-93be-461e-a382-512e1ef54f8a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.164880] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26148782-27b9-4253-bf7e-09862eff785c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.175382] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55f4d808-ca90-4d8d-bb65-1b844fabf7cd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.207686] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda37929-8a3a-47ca-b904-cfbe40f85513 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.211358] env[61970]: DEBUG nova.network.neutron [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updating instance_info_cache with network_info: [{"id": "8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d", "address": "fa:16:3e:48:95:bf", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d2476c8-b7", "ovs_interfaceid": "8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.217710] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc71d65-63b6-4698-a15b-e6e1b68443a6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.231970] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.600576] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Creating linked-clone VM from snapshot {{(pid=61970) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1114.600994] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-dfff6028-03a5-40e5-8c4c-2aa98dc1c13c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.612336] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1114.612336] env[61970]: value = "task-1356199" [ 1114.612336] env[61970]: _type = "Task" [ 1114.612336] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.624491] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356199, 'name': CloneVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.714243] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "refresh_cache-a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.735309] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1114.904536] env[61970]: DEBUG oslo_concurrency.lockutils [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "5a8cdb36-1845-4806-a6bc-22afb47886e2" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.904900] env[61970]: DEBUG oslo_concurrency.lockutils [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "5a8cdb36-1845-4806-a6bc-22afb47886e2" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.904966] env[61970]: INFO nova.compute.manager [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Attaching volume 9cb12dfc-bdee-43a7-b4ba-16190814effd to /dev/sdb [ 1114.938731] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c0a517-5913-44ba-960a-57282383dfaa {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.946529] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a098b483-c2d3-4b05-98a2-ce3957de13c9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.962823] env[61970]: DEBUG nova.virt.block_device [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating existing volume attachment record: 86a1bdc7-8f44-4200-9078-a37c3b3700b9 {{(pid=61970) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1115.123536] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356199, 'name': CloneVM_Task} progress is 94%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.239665] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1115.239955] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.226s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.624147] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356199, 'name': CloneVM_Task} progress is 95%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.124264] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356199, 'name': CloneVM_Task, 'duration_secs': 1.327375} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.124558] env[61970]: INFO nova.virt.vmwareapi.vmops [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Created linked-clone VM from snapshot [ 1116.125280] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f3a140-fd24-4dcb-af3b-b7a09638d29a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.132299] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Uploading image e7e95577-cfb6-4cb0-9088-e43964fd7b75 {{(pid=61970) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1116.157936] env[61970]: DEBUG oslo_vmware.rw_handles [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1116.157936] env[61970]: value = "vm-288888" [ 1116.157936] env[61970]: _type = "VirtualMachine" [ 1116.157936] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1116.158246] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5e15ee0b-3fe7-4b09-82f6-80eb07cec8df {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.164993] env[61970]: DEBUG oslo_vmware.rw_handles [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lease: (returnval){ [ 1116.164993] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5260c3c2-bb2d-4a6f-4e7e-0f0c7d3404e6" [ 1116.164993] env[61970]: _type = "HttpNfcLease" [ 1116.164993] env[61970]: } obtained for exporting VM: (result){ [ 1116.164993] env[61970]: value = "vm-288888" [ 1116.164993] env[61970]: _type = "VirtualMachine" [ 1116.164993] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1116.165344] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the lease: (returnval){ [ 1116.165344] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5260c3c2-bb2d-4a6f-4e7e-0f0c7d3404e6" [ 1116.165344] env[61970]: _type = "HttpNfcLease" [ 1116.165344] env[61970]: } to be ready. {{(pid=61970) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1116.171645] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1116.171645] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5260c3c2-bb2d-4a6f-4e7e-0f0c7d3404e6" [ 1116.171645] env[61970]: _type = "HttpNfcLease" [ 1116.171645] env[61970]: } is initializing. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1116.231137] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e153ad-b0ba-48d4-b38b-caa6f1d441f3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.249393] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updating instance 'a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8' progress to 0 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1116.674504] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1116.674504] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5260c3c2-bb2d-4a6f-4e7e-0f0c7d3404e6" [ 1116.674504] env[61970]: _type = "HttpNfcLease" [ 1116.674504] env[61970]: } is ready. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1116.674799] env[61970]: DEBUG oslo_vmware.rw_handles [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1116.674799] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5260c3c2-bb2d-4a6f-4e7e-0f0c7d3404e6" [ 1116.674799] env[61970]: _type = "HttpNfcLease" [ 1116.674799] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1116.675512] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c71d4234-87a3-40a5-8731-84d1f5f73022 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.682273] env[61970]: DEBUG oslo_vmware.rw_handles [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d0fd99-acf3-05c0-2186-5e7f0370abb3/disk-0.vmdk from lease info. {{(pid=61970) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1116.682451] env[61970]: DEBUG oslo_vmware.rw_handles [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d0fd99-acf3-05c0-2186-5e7f0370abb3/disk-0.vmdk for reading. {{(pid=61970) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1116.754980] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1116.755290] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6b9f8947-d71d-4397-b68c-eb72593d449e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.762392] env[61970]: DEBUG oslo_vmware.api [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1116.762392] env[61970]: value = "task-1356205" [ 1116.762392] env[61970]: _type = "Task" [ 1116.762392] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.769734] env[61970]: DEBUG oslo_vmware.api [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356205, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.772085] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4696cc4b-c5f7-4692-82c0-b40733d81b06 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.272207] env[61970]: DEBUG oslo_vmware.api [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356205, 'name': PowerOffVM_Task, 'duration_secs': 0.205623} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.272535] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1117.272624] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updating instance 'a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8' progress to 17 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1117.723075] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Volume attach. Driver type: vmdk {{(pid=61970) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1117.723502] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288886', 'volume_id': 'e4b113f3-7ebc-465c-95d5-8d14ceaca257', 'name': 'volume-e4b113f3-7ebc-465c-95d5-8d14ceaca257', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd96208cf-fcc5-4014-ac23-1d56ea97050a', 'attached_at': '', 'detached_at': '', 'volume_id': 'e4b113f3-7ebc-465c-95d5-8d14ceaca257', 'serial': 'e4b113f3-7ebc-465c-95d5-8d14ceaca257'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1117.724846] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30298bd5-0a34-4409-94a5-4b6ea1c90ed6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.750605] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612a1312-4076-4127-ab55-73ead82386e0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.786087] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1117.786446] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1117.786673] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1117.786875] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1117.787105] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1117.787313] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1117.787535] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1117.787754] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1117.787937] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1117.788196] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1117.788432] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1117.800499] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] volume-e4b113f3-7ebc-465c-95d5-8d14ceaca257/volume-e4b113f3-7ebc-465c-95d5-8d14ceaca257.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1117.800871] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86afd73a-6f5e-47ce-9eb1-6703e9c049d7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.810960] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f59fc27a-aa0b-45e6-aba4-3429158a2ccf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.830182] env[61970]: DEBUG oslo_vmware.api [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1117.830182] env[61970]: value = "task-1356208" [ 1117.830182] env[61970]: _type = "Task" [ 1117.830182] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.831929] env[61970]: DEBUG oslo_vmware.api [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1117.831929] env[61970]: value = "task-1356207" [ 1117.831929] env[61970]: _type = "Task" [ 1117.831929] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.843283] env[61970]: DEBUG oslo_vmware.api [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356208, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.846690] env[61970]: DEBUG oslo_vmware.api [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356207, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.344833] env[61970]: DEBUG oslo_vmware.api [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356208, 'name': ReconfigVM_Task, 'duration_secs': 0.177473} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.348797] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updating instance 'a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8' progress to 33 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1118.353077] env[61970]: DEBUG oslo_vmware.api [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356207, 'name': ReconfigVM_Task, 'duration_secs': 0.397641} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.353588] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Reconfigured VM instance instance-00000064 to attach disk [datastore1] volume-e4b113f3-7ebc-465c-95d5-8d14ceaca257/volume-e4b113f3-7ebc-465c-95d5-8d14ceaca257.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1118.358331] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a0ef7ee-d594-44fb-896d-fdaa81200829 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.373894] env[61970]: DEBUG oslo_vmware.api [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1118.373894] env[61970]: value = "task-1356209" [ 1118.373894] env[61970]: _type = "Task" [ 1118.373894] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.382456] env[61970]: DEBUG oslo_vmware.api [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356209, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.855780] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1118.856155] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1118.856347] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1118.856689] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1118.856923] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1118.857128] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1118.857387] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1118.857594] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1118.857806] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1118.858060] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1118.858310] env[61970]: DEBUG nova.virt.hardware [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1118.863794] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Reconfiguring VM instance instance-0000006a to detach disk 2000 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1118.864170] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-563d9525-9e5f-47d8-95ea-74385c6428c1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.885496] env[61970]: DEBUG oslo_vmware.api [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356209, 'name': ReconfigVM_Task, 'duration_secs': 0.14953} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.886772] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288886', 'volume_id': 'e4b113f3-7ebc-465c-95d5-8d14ceaca257', 'name': 'volume-e4b113f3-7ebc-465c-95d5-8d14ceaca257', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd96208cf-fcc5-4014-ac23-1d56ea97050a', 'attached_at': '', 'detached_at': '', 'volume_id': 'e4b113f3-7ebc-465c-95d5-8d14ceaca257', 'serial': 'e4b113f3-7ebc-465c-95d5-8d14ceaca257'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1118.888304] env[61970]: DEBUG oslo_vmware.api [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1118.888304] env[61970]: value = "task-1356210" [ 1118.888304] env[61970]: _type = "Task" [ 1118.888304] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.896041] env[61970]: DEBUG oslo_vmware.api [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356210, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.402716] env[61970]: DEBUG oslo_vmware.api [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356210, 'name': ReconfigVM_Task, 'duration_secs': 0.168377} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.403117] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Reconfigured VM instance instance-0000006a to detach disk 2000 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1119.404597] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1170b244-be26-44ca-aacc-5ae46798f0dc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.429859] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] volume-ed3d65b8-03da-4484-8a98-b0e8211e6d54/volume-ed3d65b8-03da-4484-8a98-b0e8211e6d54.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1119.432336] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e761bc55-1a1c-4dd0-8657-a7ab42b8a61f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.452560] env[61970]: DEBUG oslo_vmware.api [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1119.452560] env[61970]: value = "task-1356211" [ 1119.452560] env[61970]: _type = "Task" [ 1119.452560] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.461518] env[61970]: DEBUG oslo_vmware.api [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356211, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.512761] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Volume attach. Driver type: vmdk {{(pid=61970) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1119.513014] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288890', 'volume_id': '9cb12dfc-bdee-43a7-b4ba-16190814effd', 'name': 'volume-9cb12dfc-bdee-43a7-b4ba-16190814effd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5a8cdb36-1845-4806-a6bc-22afb47886e2', 'attached_at': '', 'detached_at': '', 'volume_id': '9cb12dfc-bdee-43a7-b4ba-16190814effd', 'serial': '9cb12dfc-bdee-43a7-b4ba-16190814effd'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1119.513954] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e5588d-3d51-4b8f-9fdc-47c518d1d724 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.531014] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b544b2-49ce-4f12-8d4d-9065f7691b4c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.555818] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] volume-9cb12dfc-bdee-43a7-b4ba-16190814effd/volume-9cb12dfc-bdee-43a7-b4ba-16190814effd.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1119.556129] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2efa6948-f345-40c1-b977-00ccb63f6c86 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.574492] env[61970]: DEBUG oslo_vmware.api [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1119.574492] env[61970]: value = "task-1356212" [ 1119.574492] env[61970]: _type = "Task" [ 1119.574492] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.582881] env[61970]: DEBUG oslo_vmware.api [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356212, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.947307] env[61970]: DEBUG nova.objects.instance [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lazy-loading 'flavor' on Instance uuid d96208cf-fcc5-4014-ac23-1d56ea97050a {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1119.963500] env[61970]: DEBUG oslo_vmware.api [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356211, 'name': ReconfigVM_Task, 'duration_secs': 0.264323} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.964567] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Reconfigured VM instance instance-0000006a to attach disk [datastore1] volume-ed3d65b8-03da-4484-8a98-b0e8211e6d54/volume-ed3d65b8-03da-4484-8a98-b0e8211e6d54.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1119.964868] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updating instance 'a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8' progress to 50 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1120.085111] env[61970]: DEBUG oslo_vmware.api [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356212, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.453253] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a33aeff1-1d53-448e-9b4c-ffda4dd94578 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.326s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.472340] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9b63dd-86a0-42a5-9d9b-3df70a87cec9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.492539] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a86ba43-c18b-47c2-aee8-eead8c6f61d6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.510990] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updating instance 'a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8' progress to 67 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1120.586871] env[61970]: DEBUG oslo_vmware.api [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356212, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.657009] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60dc5c33-4078-465a-91c5-63e05cbdcf33 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.657009] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60dc5c33-4078-465a-91c5-63e05cbdcf33 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.086052] env[61970]: DEBUG oslo_vmware.api [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356212, 'name': ReconfigVM_Task, 'duration_secs': 1.4705} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.086875] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Reconfigured VM instance instance-00000068 to attach disk [datastore1] volume-9cb12dfc-bdee-43a7-b4ba-16190814effd/volume-9cb12dfc-bdee-43a7-b4ba-16190814effd.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1121.092221] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a2a4f50-221f-4a8a-92f3-0d6602c7a494 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.107263] env[61970]: DEBUG oslo_vmware.api [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1121.107263] env[61970]: value = "task-1356213" [ 1121.107263] env[61970]: _type = "Task" [ 1121.107263] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.117592] env[61970]: DEBUG oslo_vmware.api [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356213, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.159677] env[61970]: INFO nova.compute.manager [None req-60dc5c33-4078-465a-91c5-63e05cbdcf33 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Detaching volume 3ff6c6bf-861a-44c7-8436-f62c37e40dfb [ 1121.192055] env[61970]: INFO nova.virt.block_device [None req-60dc5c33-4078-465a-91c5-63e05cbdcf33 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Attempting to driver detach volume 3ff6c6bf-861a-44c7-8436-f62c37e40dfb from mountpoint /dev/sdb [ 1121.192055] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-60dc5c33-4078-465a-91c5-63e05cbdcf33 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Volume detach. Driver type: vmdk {{(pid=61970) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1121.192055] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-60dc5c33-4078-465a-91c5-63e05cbdcf33 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288884', 'volume_id': '3ff6c6bf-861a-44c7-8436-f62c37e40dfb', 'name': 'volume-3ff6c6bf-861a-44c7-8436-f62c37e40dfb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd96208cf-fcc5-4014-ac23-1d56ea97050a', 'attached_at': '', 'detached_at': '', 'volume_id': '3ff6c6bf-861a-44c7-8436-f62c37e40dfb', 'serial': '3ff6c6bf-861a-44c7-8436-f62c37e40dfb'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1121.192789] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f1f539e-9766-4f8b-9904-c1de12326975 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.219619] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1084c3d4-9e5b-4a24-9a9b-3dc3b8ac3097 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.226432] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84051c00-9f8b-4898-b292-560495339cd9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.249538] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde4b61e-5e9d-4ca9-85fb-dd18070005ae {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.264577] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-60dc5c33-4078-465a-91c5-63e05cbdcf33 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] The volume has not been displaced from its original location: [datastore1] volume-3ff6c6bf-861a-44c7-8436-f62c37e40dfb/volume-3ff6c6bf-861a-44c7-8436-f62c37e40dfb.vmdk. No consolidation needed. {{(pid=61970) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1121.269835] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-60dc5c33-4078-465a-91c5-63e05cbdcf33 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Reconfiguring VM instance instance-00000064 to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1121.270207] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3354964a-0749-4500-8107-e987e2c52b13 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.287324] env[61970]: DEBUG oslo_vmware.api [None req-60dc5c33-4078-465a-91c5-63e05cbdcf33 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1121.287324] env[61970]: value = "task-1356214" [ 1121.287324] env[61970]: _type = "Task" [ 1121.287324] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.294971] env[61970]: DEBUG oslo_vmware.api [None req-60dc5c33-4078-465a-91c5-63e05cbdcf33 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356214, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.616670] env[61970]: DEBUG oslo_vmware.api [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356213, 'name': ReconfigVM_Task, 'duration_secs': 0.156969} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.618831] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288890', 'volume_id': '9cb12dfc-bdee-43a7-b4ba-16190814effd', 'name': 'volume-9cb12dfc-bdee-43a7-b4ba-16190814effd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5a8cdb36-1845-4806-a6bc-22afb47886e2', 'attached_at': '', 'detached_at': '', 'volume_id': '9cb12dfc-bdee-43a7-b4ba-16190814effd', 'serial': '9cb12dfc-bdee-43a7-b4ba-16190814effd'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1121.796907] env[61970]: DEBUG oslo_vmware.api [None req-60dc5c33-4078-465a-91c5-63e05cbdcf33 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356214, 'name': ReconfigVM_Task, 'duration_secs': 0.233762} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.797225] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-60dc5c33-4078-465a-91c5-63e05cbdcf33 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Reconfigured VM instance instance-00000064 to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1121.801926] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f14fcec-e346-4c0c-802c-df0c7a8fbb0a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.816568] env[61970]: DEBUG oslo_vmware.api [None req-60dc5c33-4078-465a-91c5-63e05cbdcf33 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1121.816568] env[61970]: value = "task-1356215" [ 1121.816568] env[61970]: _type = "Task" [ 1121.816568] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.824970] env[61970]: DEBUG oslo_vmware.api [None req-60dc5c33-4078-465a-91c5-63e05cbdcf33 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356215, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.158058] env[61970]: DEBUG nova.network.neutron [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Port 8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d binding to destination host cpu-1 is already ACTIVE {{(pid=61970) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1122.327434] env[61970]: DEBUG oslo_vmware.api [None req-60dc5c33-4078-465a-91c5-63e05cbdcf33 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356215, 'name': ReconfigVM_Task, 'duration_secs': 0.143253} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.327774] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-60dc5c33-4078-465a-91c5-63e05cbdcf33 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288884', 'volume_id': '3ff6c6bf-861a-44c7-8436-f62c37e40dfb', 'name': 'volume-3ff6c6bf-861a-44c7-8436-f62c37e40dfb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd96208cf-fcc5-4014-ac23-1d56ea97050a', 'attached_at': '', 'detached_at': '', 'volume_id': '3ff6c6bf-861a-44c7-8436-f62c37e40dfb', 'serial': '3ff6c6bf-861a-44c7-8436-f62c37e40dfb'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1122.658083] env[61970]: DEBUG nova.objects.instance [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lazy-loading 'flavor' on Instance uuid 5a8cdb36-1845-4806-a6bc-22afb47886e2 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1122.870159] env[61970]: DEBUG nova.objects.instance [None req-60dc5c33-4078-465a-91c5-63e05cbdcf33 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lazy-loading 'flavor' on Instance uuid d96208cf-fcc5-4014-ac23-1d56ea97050a {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1123.163363] env[61970]: DEBUG oslo_concurrency.lockutils [None req-81ed1165-d660-4d5d-b03f-c5f0e31abb9f tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "5a8cdb36-1845-4806-a6bc-22afb47886e2" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.258s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.179535] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.179785] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.180010] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.878345] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60dc5c33-4078-465a-91c5-63e05cbdcf33 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.221s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.925166] env[61970]: DEBUG oslo_concurrency.lockutils [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.925529] env[61970]: DEBUG oslo_concurrency.lockutils [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.972836] env[61970]: DEBUG oslo_vmware.rw_handles [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d0fd99-acf3-05c0-2186-5e7f0370abb3/disk-0.vmdk. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1123.973795] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e519fb20-063a-46f8-9455-4a79d37cc026 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.980320] env[61970]: DEBUG oslo_vmware.rw_handles [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d0fd99-acf3-05c0-2186-5e7f0370abb3/disk-0.vmdk is in state: ready. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1123.980522] env[61970]: ERROR oslo_vmware.rw_handles [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d0fd99-acf3-05c0-2186-5e7f0370abb3/disk-0.vmdk due to incomplete transfer. [ 1123.980748] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-84546e36-9ccf-4eb1-ad55-93a84e94b0c0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.987277] env[61970]: DEBUG oslo_vmware.rw_handles [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d0fd99-acf3-05c0-2186-5e7f0370abb3/disk-0.vmdk. {{(pid=61970) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1123.987506] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Uploaded image e7e95577-cfb6-4cb0-9088-e43964fd7b75 to the Glance image server {{(pid=61970) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1123.989957] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Destroying the VM {{(pid=61970) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1123.990219] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-eca75e32-3f90-4c5a-9945-f4fda830eece {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.995850] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1123.995850] env[61970]: value = "task-1356216" [ 1123.995850] env[61970]: _type = "Task" [ 1123.995850] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.003655] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356216, 'name': Destroy_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.233406] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "refresh_cache-a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.233720] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "refresh_cache-a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.234007] env[61970]: DEBUG nova.network.neutron [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1124.300139] env[61970]: DEBUG nova.compute.manager [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Stashing vm_state: active {{(pid=61970) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1124.428288] env[61970]: INFO nova.compute.manager [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Detaching volume e4b113f3-7ebc-465c-95d5-8d14ceaca257 [ 1124.461611] env[61970]: INFO nova.virt.block_device [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Attempting to driver detach volume e4b113f3-7ebc-465c-95d5-8d14ceaca257 from mountpoint /dev/sdc [ 1124.461853] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Volume detach. Driver type: vmdk {{(pid=61970) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1124.462057] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288886', 'volume_id': 'e4b113f3-7ebc-465c-95d5-8d14ceaca257', 'name': 'volume-e4b113f3-7ebc-465c-95d5-8d14ceaca257', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd96208cf-fcc5-4014-ac23-1d56ea97050a', 'attached_at': '', 'detached_at': '', 'volume_id': 'e4b113f3-7ebc-465c-95d5-8d14ceaca257', 'serial': 'e4b113f3-7ebc-465c-95d5-8d14ceaca257'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1124.462950] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf5a016-2176-410b-bc19-67c54628c2d5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.485253] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f85c0c5-5acc-42bf-8ad4-1afad96b9410 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.492719] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd39909-7782-4740-ae92-22dea6ab2157 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.514738] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2462582f-3df4-4f84-b5b0-1dae9ecd5e51 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.519842] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356216, 'name': Destroy_Task, 'duration_secs': 0.304213} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.520441] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Destroyed the VM [ 1124.520691] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Deleting Snapshot of the VM instance {{(pid=61970) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1124.520921] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-14294a51-dc4e-449b-9cbd-7d9a31665292 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.532369] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] The volume has not been displaced from its original location: [datastore1] volume-e4b113f3-7ebc-465c-95d5-8d14ceaca257/volume-e4b113f3-7ebc-465c-95d5-8d14ceaca257.vmdk. No consolidation needed. {{(pid=61970) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1124.537593] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Reconfiguring VM instance instance-00000064 to detach disk 2002 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1124.538245] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7efa0361-a7dd-46e8-b6f3-2bdde8efa106 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.552244] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1124.552244] env[61970]: value = "task-1356217" [ 1124.552244] env[61970]: _type = "Task" [ 1124.552244] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.557941] env[61970]: DEBUG oslo_vmware.api [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1124.557941] env[61970]: value = "task-1356218" [ 1124.557941] env[61970]: _type = "Task" [ 1124.557941] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.563625] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356217, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.569777] env[61970]: DEBUG oslo_vmware.api [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356218, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.628173] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.821452] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.821791] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.957300] env[61970]: DEBUG nova.network.neutron [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updating instance_info_cache with network_info: [{"id": "8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d", "address": "fa:16:3e:48:95:bf", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d2476c8-b7", "ovs_interfaceid": "8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.066209] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356217, 'name': RemoveSnapshot_Task, 'duration_secs': 0.360083} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.066855] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Deleted Snapshot of the VM instance {{(pid=61970) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1125.067182] env[61970]: DEBUG nova.compute.manager [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1125.067937] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb2d4ab-2ff8-48d2-9a56-26edd0ecdd9e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.073114] env[61970]: DEBUG oslo_vmware.api [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356218, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.132677] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Getting list of instances from cluster (obj){ [ 1125.132677] env[61970]: value = "domain-c8" [ 1125.132677] env[61970]: _type = "ClusterComputeResource" [ 1125.132677] env[61970]: } {{(pid=61970) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1125.133740] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c30db4f-8942-41d2-90b5-e2526eb1ebf7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.151930] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Got total of 7 instances {{(pid=61970) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1125.152101] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Triggering sync for uuid 0304640a-6c46-4f57-a8b8-5058da2e3de4 {{(pid=61970) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1125.152294] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Triggering sync for uuid d96208cf-fcc5-4014-ac23-1d56ea97050a {{(pid=61970) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1125.152447] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Triggering sync for uuid 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6 {{(pid=61970) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1125.152594] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Triggering sync for uuid 183133c6-7e4e-4dd4-b8e0-cb4f8338264c {{(pid=61970) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1125.152742] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Triggering sync for uuid 5a8cdb36-1845-4806-a6bc-22afb47886e2 {{(pid=61970) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1125.152888] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Triggering sync for uuid df0a2338-07e0-4148-8729-a290b88207b3 {{(pid=61970) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1125.153044] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Triggering sync for uuid a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8 {{(pid=61970) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1125.153393] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "0304640a-6c46-4f57-a8b8-5058da2e3de4" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.153611] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "0304640a-6c46-4f57-a8b8-5058da2e3de4" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.153869] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.154131] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.154347] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "183133c6-7e4e-4dd4-b8e0-cb4f8338264c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.154531] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "183133c6-7e4e-4dd4-b8e0-cb4f8338264c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.154766] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "5a8cdb36-1845-4806-a6bc-22afb47886e2" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.154946] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "5a8cdb36-1845-4806-a6bc-22afb47886e2" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.155131] env[61970]: INFO nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] During sync_power_state the instance has a pending task (resize_prep). Skip. [ 1125.155295] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "5a8cdb36-1845-4806-a6bc-22afb47886e2" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.155484] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "df0a2338-07e0-4148-8729-a290b88207b3" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.155663] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "df0a2338-07e0-4148-8729-a290b88207b3" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.156084] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.156084] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.156242] env[61970]: INFO nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] During sync_power_state the instance has a pending task (resize_migrated). Skip. [ 1125.156400] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.157185] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b91b3d8-12e5-4639-b204-976c03ad0206 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.159986] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d3fb646-3f3c-40ec-a83d-fbe123e24b07 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.162736] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e062bf46-1674-42a4-8c51-1372c6a20ece {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.326529] env[61970]: INFO nova.compute.claims [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1125.461505] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "refresh_cache-a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.568516] env[61970]: DEBUG oslo_vmware.api [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356218, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.582440] env[61970]: INFO nova.compute.manager [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Shelve offloading [ 1125.583919] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1125.584179] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c11503f2-5d04-4320-b497-6f8eb7c2acf3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.591611] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1125.591611] env[61970]: value = "task-1356219" [ 1125.591611] env[61970]: _type = "Task" [ 1125.591611] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.598758] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356219, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.674065] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "0304640a-6c46-4f57-a8b8-5058da2e3de4" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.520s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.674457] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "df0a2338-07e0-4148-8729-a290b88207b3" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.519s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.674775] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "183133c6-7e4e-4dd4-b8e0-cb4f8338264c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.520s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.832812] env[61970]: INFO nova.compute.resource_tracker [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating resource usage from migration 835beeae-86b7-4c24-85fe-065e481c60a0 [ 1125.932041] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e650bb-bf02-452a-b04f-7a9224b97e9c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.939330] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37d7cc2-48d3-40a5-a6c4-58c7db4bcb06 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.975884] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9a6218-c293-4a7e-bad5-317195cc615f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.980244] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76f6068-5099-4389-a7b5-df57826b57e8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.990426] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-511f266c-b50e-4cf1-8a37-0bd6d5a02ed0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.994927] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9290c238-eb1f-4deb-9489-b74872b0444b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.013266] env[61970]: DEBUG nova.compute.provider_tree [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1126.067971] env[61970]: DEBUG oslo_vmware.api [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356218, 'name': ReconfigVM_Task, 'duration_secs': 1.240989} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.068275] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Reconfigured VM instance instance-00000064 to detach disk 2002 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1126.073022] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dbf13d64-ab8a-4e4b-96db-50d6397953ee {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.087092] env[61970]: DEBUG oslo_vmware.api [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1126.087092] env[61970]: value = "task-1356220" [ 1126.087092] env[61970]: _type = "Task" [ 1126.087092] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.095348] env[61970]: DEBUG oslo_vmware.api [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356220, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.102614] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] VM already powered off {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1126.102804] env[61970]: DEBUG nova.compute.manager [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1126.103542] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e92f018-2c38-4b49-b7ae-aebe211567ad {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.109557] env[61970]: DEBUG oslo_concurrency.lockutils [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.109689] env[61970]: DEBUG oslo_concurrency.lockutils [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquired lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.110167] env[61970]: DEBUG nova.network.neutron [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1126.523812] env[61970]: DEBUG nova.scheduler.client.report [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1126.602644] env[61970]: DEBUG oslo_vmware.api [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356220, 'name': ReconfigVM_Task, 'duration_secs': 0.142434} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.603099] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288886', 'volume_id': 'e4b113f3-7ebc-465c-95d5-8d14ceaca257', 'name': 'volume-e4b113f3-7ebc-465c-95d5-8d14ceaca257', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd96208cf-fcc5-4014-ac23-1d56ea97050a', 'attached_at': '', 'detached_at': '', 'volume_id': 'e4b113f3-7ebc-465c-95d5-8d14ceaca257', 'serial': 'e4b113f3-7ebc-465c-95d5-8d14ceaca257'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1126.816326] env[61970]: DEBUG nova.network.neutron [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Updating instance_info_cache with network_info: [{"id": "2fa02dfe-2cb0-4a72-bc30-4f991339dcb9", "address": "fa:16:3e:df:19:94", "network": {"id": "70081995-3d6c-466e-b811-1c8e532e552f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-641102750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1d4724cbb0841859369af01f2c03db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aaf1b231-3660-4453-b4f3-44d825b9a5dd", "external-id": "nsx-vlan-transportzone-6", "segmentation_id": 6, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fa02dfe-2c", "ovs_interfaceid": "2fa02dfe-2cb0-4a72-bc30-4f991339dcb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.028827] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.207s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.029222] env[61970]: INFO nova.compute.manager [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Migrating [ 1127.093480] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7255bf00-38ba-44f7-b2d3-dc010374c76f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.112349] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d7f1fd-d35a-48b8-9a41-e3953c60cbdb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.119172] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updating instance 'a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8' progress to 83 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1127.143291] env[61970]: DEBUG nova.objects.instance [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lazy-loading 'flavor' on Instance uuid d96208cf-fcc5-4014-ac23-1d56ea97050a {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1127.318997] env[61970]: DEBUG oslo_concurrency.lockutils [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Releasing lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.544143] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.544364] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.544553] env[61970]: DEBUG nova.network.neutron [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1127.550942] env[61970]: DEBUG nova.compute.manager [req-4120b2f2-42b3-4718-b956-5f44a8f9ac04 req-62732179-8b04-4d3a-8d7b-0535f8c70bf0 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Received event network-vif-unplugged-2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1127.552114] env[61970]: DEBUG oslo_concurrency.lockutils [req-4120b2f2-42b3-4718-b956-5f44a8f9ac04 req-62732179-8b04-4d3a-8d7b-0535f8c70bf0 service nova] Acquiring lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.552114] env[61970]: DEBUG oslo_concurrency.lockutils [req-4120b2f2-42b3-4718-b956-5f44a8f9ac04 req-62732179-8b04-4d3a-8d7b-0535f8c70bf0 service nova] Lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.552114] env[61970]: DEBUG oslo_concurrency.lockutils [req-4120b2f2-42b3-4718-b956-5f44a8f9ac04 req-62732179-8b04-4d3a-8d7b-0535f8c70bf0 service nova] Lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.552114] env[61970]: DEBUG nova.compute.manager [req-4120b2f2-42b3-4718-b956-5f44a8f9ac04 req-62732179-8b04-4d3a-8d7b-0535f8c70bf0 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] No waiting events found dispatching network-vif-unplugged-2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1127.552114] env[61970]: WARNING nova.compute.manager [req-4120b2f2-42b3-4718-b956-5f44a8f9ac04 req-62732179-8b04-4d3a-8d7b-0535f8c70bf0 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Received unexpected event network-vif-unplugged-2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 for instance with vm_state shelved and task_state shelving_offloading. [ 1127.623969] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "df0a2338-07e0-4148-8729-a290b88207b3" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.624243] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "df0a2338-07e0-4148-8729-a290b88207b3" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.628837] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1127.631682] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b345f609-7a54-40d2-836f-8405dc518050 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.639605] env[61970]: DEBUG oslo_vmware.api [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1127.639605] env[61970]: value = "task-1356221" [ 1127.639605] env[61970]: _type = "Task" [ 1127.639605] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.650709] env[61970]: DEBUG oslo_vmware.api [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356221, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.655917] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1127.656936] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a53dfa-1edf-482b-92d5-680f42760f23 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.665398] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1127.665627] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-99c90b0d-7197-4f33-ace2-23d6038ded6a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.732590] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1127.732854] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1127.733061] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Deleting the datastore file [datastore2] 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1127.733353] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8812d693-5668-4c39-9395-179ad3dfbf3e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.739058] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1127.739058] env[61970]: value = "task-1356223" [ 1127.739058] env[61970]: _type = "Task" [ 1127.739058] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.746376] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356223, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.132681] env[61970]: DEBUG nova.compute.utils [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1128.152490] env[61970]: DEBUG oslo_vmware.api [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356221, 'name': PowerOnVM_Task, 'duration_secs': 0.409357} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.152866] env[61970]: DEBUG oslo_concurrency.lockutils [None req-232f9d9f-debb-4b4f-bb45-35bc55d20ec3 tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.227s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.153781] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1128.153977] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ef7d4980-92e3-4381-ac55-fd109bda83d3 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updating instance 'a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8' progress to 100 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1128.157418] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 3.003s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.158523] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-003016b8-95c1-4637-87ba-cd7bbd7aaea9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.248965] env[61970]: DEBUG oslo_vmware.api [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356223, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154726} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.249289] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1128.249489] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1128.249665] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1128.257795] env[61970]: DEBUG nova.network.neutron [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating instance_info_cache with network_info: [{"id": "696b0179-1272-4ae7-ab15-727fe4682c42", "address": "fa:16:3e:9e:3f:74", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap696b0179-12", "ovs_interfaceid": "696b0179-1272-4ae7-ab15-727fe4682c42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.271402] env[61970]: INFO nova.scheduler.client.report [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Deleted allocations for instance 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6 [ 1128.372782] env[61970]: DEBUG oslo_concurrency.lockutils [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.635958] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "df0a2338-07e0-4148-8729-a290b88207b3" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.668486] env[61970]: INFO nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] During sync_power_state the instance has a pending task (deleting). Skip. [ 1128.668712] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.511s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.668948] env[61970]: DEBUG oslo_concurrency.lockutils [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.296s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.669173] env[61970]: DEBUG oslo_concurrency.lockutils [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "d96208cf-fcc5-4014-ac23-1d56ea97050a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.669376] env[61970]: DEBUG oslo_concurrency.lockutils [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.669544] env[61970]: DEBUG oslo_concurrency.lockutils [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.672282] env[61970]: INFO nova.compute.manager [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Terminating instance [ 1128.673975] env[61970]: DEBUG nova.compute.manager [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1128.674452] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1128.675061] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7698950c-6456-4e40-82b0-d61914d07e4b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.682739] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1128.682739] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-56f99c9e-cf7c-42eb-9f2d-d317dcc4b6de {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.688522] env[61970]: DEBUG oslo_vmware.api [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1128.688522] env[61970]: value = "task-1356224" [ 1128.688522] env[61970]: _type = "Task" [ 1128.688522] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.697506] env[61970]: DEBUG oslo_vmware.api [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356224, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.760247] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.775440] env[61970]: DEBUG oslo_concurrency.lockutils [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.775713] env[61970]: DEBUG oslo_concurrency.lockutils [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.775942] env[61970]: DEBUG nova.objects.instance [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lazy-loading 'resources' on Instance uuid 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1129.200527] env[61970]: DEBUG oslo_vmware.api [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356224, 'name': PowerOffVM_Task, 'duration_secs': 0.175485} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.200936] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1129.201094] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1129.201249] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f208a92d-d4e6-48c0-97ea-ac25b4abd100 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.266276] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1129.266491] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1129.266673] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Deleting the datastore file [datastore1] d96208cf-fcc5-4014-ac23-1d56ea97050a {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1129.267361] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d3bee260-1a82-47ff-91b8-9c54e8a81d75 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.273620] env[61970]: DEBUG oslo_vmware.api [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for the task: (returnval){ [ 1129.273620] env[61970]: value = "task-1356226" [ 1129.273620] env[61970]: _type = "Task" [ 1129.273620] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.278246] env[61970]: DEBUG nova.objects.instance [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lazy-loading 'numa_topology' on Instance uuid 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1129.283639] env[61970]: DEBUG oslo_vmware.api [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356226, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.620129] env[61970]: DEBUG nova.compute.manager [req-77557db4-4b98-47e1-b400-04cc57e7b675 req-d11cc715-6464-4df8-a5d1-d8d09e86f478 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Received event network-changed-2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1129.620495] env[61970]: DEBUG nova.compute.manager [req-77557db4-4b98-47e1-b400-04cc57e7b675 req-d11cc715-6464-4df8-a5d1-d8d09e86f478 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Refreshing instance network info cache due to event network-changed-2fa02dfe-2cb0-4a72-bc30-4f991339dcb9. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1129.620495] env[61970]: DEBUG oslo_concurrency.lockutils [req-77557db4-4b98-47e1-b400-04cc57e7b675 req-d11cc715-6464-4df8-a5d1-d8d09e86f478 service nova] Acquiring lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.620728] env[61970]: DEBUG oslo_concurrency.lockutils [req-77557db4-4b98-47e1-b400-04cc57e7b675 req-d11cc715-6464-4df8-a5d1-d8d09e86f478 service nova] Acquired lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.620950] env[61970]: DEBUG nova.network.neutron [req-77557db4-4b98-47e1-b400-04cc57e7b675 req-d11cc715-6464-4df8-a5d1-d8d09e86f478 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Refreshing network info cache for port 2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1129.718027] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "df0a2338-07e0-4148-8729-a290b88207b3" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.718362] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "df0a2338-07e0-4148-8729-a290b88207b3" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.718612] env[61970]: INFO nova.compute.manager [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Attaching volume 6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe to /dev/sdb [ 1129.751452] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00243e86-4ce7-435b-aeb7-9729ccae83d8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.759321] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57c62e98-0b74-4316-ba3a-f37ee1d86cfd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.773490] env[61970]: DEBUG nova.virt.block_device [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Updating existing volume attachment record: a1572d40-1c42-4ed7-a3b5-8d4719b788dd {{(pid=61970) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1129.783319] env[61970]: DEBUG oslo_vmware.api [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Task: {'id': task-1356226, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142413} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.784144] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1129.784360] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1129.784537] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1129.784709] env[61970]: INFO nova.compute.manager [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1129.784944] env[61970]: DEBUG oslo.service.loopingcall [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1129.785278] env[61970]: DEBUG nova.objects.base [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Object Instance<3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6> lazy-loaded attributes: resources,numa_topology {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1129.787749] env[61970]: DEBUG nova.compute.manager [-] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1129.787854] env[61970]: DEBUG nova.network.neutron [-] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1129.896861] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3014ce9-07c0-4a67-b5d0-3cda8fe72ef5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.904822] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0a04dd4-7d5a-4345-bafe-a6b9fd74b751 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.942098] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ae29248-c20f-4946-9ce0-1a28f93f17c3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.949806] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac66a1c-3c4c-4115-ae7a-87c8d5cb88da {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.963455] env[61970]: DEBUG nova.compute.provider_tree [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1130.155976] env[61970]: DEBUG oslo_concurrency.lockutils [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.156243] env[61970]: DEBUG oslo_concurrency.lockutils [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.156434] env[61970]: DEBUG nova.compute.manager [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Going to confirm migration 5 {{(pid=61970) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1130.283164] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45034af-0814-4285-9a06-25f75f76369b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.304113] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating instance '5a8cdb36-1845-4806-a6bc-22afb47886e2' progress to 0 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1130.362024] env[61970]: DEBUG nova.network.neutron [req-77557db4-4b98-47e1-b400-04cc57e7b675 req-d11cc715-6464-4df8-a5d1-d8d09e86f478 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Updated VIF entry in instance network info cache for port 2fa02dfe-2cb0-4a72-bc30-4f991339dcb9. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1130.362422] env[61970]: DEBUG nova.network.neutron [req-77557db4-4b98-47e1-b400-04cc57e7b675 req-d11cc715-6464-4df8-a5d1-d8d09e86f478 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Updating instance_info_cache with network_info: [{"id": "2fa02dfe-2cb0-4a72-bc30-4f991339dcb9", "address": "fa:16:3e:df:19:94", "network": {"id": "70081995-3d6c-466e-b811-1c8e532e552f", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-641102750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1d4724cbb0841859369af01f2c03db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap2fa02dfe-2c", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.466366] env[61970]: DEBUG nova.scheduler.client.report [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1130.652901] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.741120] env[61970]: DEBUG oslo_concurrency.lockutils [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "refresh_cache-a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.741120] env[61970]: DEBUG oslo_concurrency.lockutils [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquired lock "refresh_cache-a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.741120] env[61970]: DEBUG nova.network.neutron [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1130.741120] env[61970]: DEBUG nova.objects.instance [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lazy-loading 'info_cache' on Instance uuid a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1130.754022] env[61970]: DEBUG nova.network.neutron [-] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.810073] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1130.810417] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-de7231bd-2af8-48b2-a14e-7880b7d1ca82 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.818394] env[61970]: DEBUG oslo_vmware.api [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1130.818394] env[61970]: value = "task-1356228" [ 1130.818394] env[61970]: _type = "Task" [ 1130.818394] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.826656] env[61970]: DEBUG oslo_vmware.api [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356228, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.864900] env[61970]: DEBUG oslo_concurrency.lockutils [req-77557db4-4b98-47e1-b400-04cc57e7b675 req-d11cc715-6464-4df8-a5d1-d8d09e86f478 service nova] Releasing lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.971986] env[61970]: DEBUG oslo_concurrency.lockutils [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.196s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.256779] env[61970]: INFO nova.compute.manager [-] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Took 1.47 seconds to deallocate network for instance. [ 1131.329279] env[61970]: DEBUG oslo_vmware.api [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356228, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.480899] env[61970]: DEBUG oslo_concurrency.lockutils [None req-29515c14-58a5-45de-b595-b7c73d4926b4 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.482s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.481856] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 6.328s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.482069] env[61970]: INFO nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] During sync_power_state the instance has a pending task (shelving_image_uploading). Skip. [ 1131.482250] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.482671] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.830s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.482858] env[61970]: INFO nova.compute.manager [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Unshelving [ 1131.642960] env[61970]: DEBUG nova.compute.manager [req-0962ed74-31c9-47ea-a308-e2a32899a3e9 req-97f40fc1-aba9-42e4-a30b-6272e85f2e6e service nova] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Received event network-vif-deleted-a7b3f775-a79c-488c-bec7-e8a2b425c91f {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1131.763410] env[61970]: DEBUG oslo_concurrency.lockutils [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.763688] env[61970]: DEBUG oslo_concurrency.lockutils [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.763909] env[61970]: DEBUG nova.objects.instance [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lazy-loading 'resources' on Instance uuid d96208cf-fcc5-4014-ac23-1d56ea97050a {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1131.828894] env[61970]: DEBUG oslo_vmware.api [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356228, 'name': PowerOffVM_Task, 'duration_secs': 0.517302} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.829221] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1131.829418] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating instance '5a8cdb36-1845-4806-a6bc-22afb47886e2' progress to 17 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1132.030233] env[61970]: DEBUG nova.network.neutron [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updating instance_info_cache with network_info: [{"id": "8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d", "address": "fa:16:3e:48:95:bf", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d2476c8-b7", "ovs_interfaceid": "8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.336105] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1132.336479] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1132.336716] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1132.337012] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1132.337247] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1132.337469] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1132.337751] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1132.338069] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1132.338257] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1132.338509] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1132.338759] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1132.346139] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc074718-970f-4060-b967-e231821b0e24 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.362268] env[61970]: DEBUG oslo_vmware.api [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1132.362268] env[61970]: value = "task-1356230" [ 1132.362268] env[61970]: _type = "Task" [ 1132.362268] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.370442] env[61970]: DEBUG oslo_vmware.api [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356230, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.400046] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f492c26-4caf-441e-a6f6-037a47138fa8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.406934] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b3df25-ab8f-4c77-a1b0-c7688096f829 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.436624] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22e70a3-a333-4e3a-b9be-fcf76d4dfea3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.443181] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd2240c-a4c4-4ceb-8767-e5a16029946b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.455665] env[61970]: DEBUG nova.compute.provider_tree [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1132.505873] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.534138] env[61970]: DEBUG oslo_concurrency.lockutils [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Releasing lock "refresh_cache-a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.534138] env[61970]: DEBUG nova.objects.instance [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lazy-loading 'migration_context' on Instance uuid a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1132.871777] env[61970]: DEBUG oslo_vmware.api [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356230, 'name': ReconfigVM_Task, 'duration_secs': 0.163691} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.872899] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating instance '5a8cdb36-1845-4806-a6bc-22afb47886e2' progress to 33 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1132.958872] env[61970]: DEBUG nova.scheduler.client.report [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1133.036264] env[61970]: DEBUG nova.objects.base [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1133.037249] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31ff3f0-fba3-48f3-b4c5-a4d1cd557557 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.057443] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d331c24-910b-4683-bd1c-f954babe2c37 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.062976] env[61970]: DEBUG oslo_vmware.api [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1133.062976] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5201d16e-eabd-5d18-0006-af448a08e7bd" [ 1133.062976] env[61970]: _type = "Task" [ 1133.062976] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.070610] env[61970]: DEBUG oslo_vmware.api [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5201d16e-eabd-5d18-0006-af448a08e7bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.378489] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1133.378815] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1133.378943] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1133.379194] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1133.379444] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1133.379649] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1133.379969] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1133.380260] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1133.380505] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1133.380695] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1133.380903] env[61970]: DEBUG nova.virt.hardware [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1133.386455] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Reconfiguring VM instance instance-00000068 to detach disk 2000 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1133.386763] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d62cf80-4564-4a23-b47f-34cc7c928b64 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.407181] env[61970]: DEBUG oslo_vmware.api [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1133.407181] env[61970]: value = "task-1356231" [ 1133.407181] env[61970]: _type = "Task" [ 1133.407181] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.416650] env[61970]: DEBUG oslo_vmware.api [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356231, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.463817] env[61970]: DEBUG oslo_concurrency.lockutils [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.700s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.467219] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.960s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.467219] env[61970]: DEBUG nova.objects.instance [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lazy-loading 'pci_requests' on Instance uuid 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1133.485296] env[61970]: INFO nova.scheduler.client.report [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Deleted allocations for instance d96208cf-fcc5-4014-ac23-1d56ea97050a [ 1133.572727] env[61970]: DEBUG oslo_vmware.api [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5201d16e-eabd-5d18-0006-af448a08e7bd, 'name': SearchDatastore_Task, 'duration_secs': 0.006674} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.573393] env[61970]: DEBUG oslo_concurrency.lockutils [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.916545] env[61970]: DEBUG oslo_vmware.api [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356231, 'name': ReconfigVM_Task, 'duration_secs': 0.198478} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.916849] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Reconfigured VM instance instance-00000068 to detach disk 2000 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1133.917644] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eabd096-33c6-425e-be3c-e56f7f134342 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.942013] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 5a8cdb36-1845-4806-a6bc-22afb47886e2/5a8cdb36-1845-4806-a6bc-22afb47886e2.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1133.942324] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af55c6bf-1b90-40c0-b605-ae714906678e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.959905] env[61970]: DEBUG oslo_vmware.api [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1133.959905] env[61970]: value = "task-1356233" [ 1133.959905] env[61970]: _type = "Task" [ 1133.959905] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.967774] env[61970]: DEBUG oslo_vmware.api [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356233, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.970483] env[61970]: DEBUG nova.objects.instance [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lazy-loading 'numa_topology' on Instance uuid 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1133.991862] env[61970]: DEBUG oslo_concurrency.lockutils [None req-efc9889a-58d3-4db1-9dbd-d93b96578b5e tempest-AttachVolumeTestJSON-1565769631 tempest-AttachVolumeTestJSON-1565769631-project-member] Lock "d96208cf-fcc5-4014-ac23-1d56ea97050a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.323s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.320946] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Volume attach. Driver type: vmdk {{(pid=61970) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1134.321216] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288891', 'volume_id': '6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe', 'name': 'volume-6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'df0a2338-07e0-4148-8729-a290b88207b3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe', 'serial': '6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1134.322241] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa5bcab-271f-47ca-937b-a2d644f0980e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.339333] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157939d8-67c7-408c-a19b-e6e469ac4407 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.367046] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] volume-6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe/volume-6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1134.367046] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae7efeff-71c4-4214-90d5-1477e9f24cc4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.385296] env[61970]: DEBUG oslo_vmware.api [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1134.385296] env[61970]: value = "task-1356234" [ 1134.385296] env[61970]: _type = "Task" [ 1134.385296] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.395776] env[61970]: DEBUG oslo_vmware.api [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356234, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.470067] env[61970]: DEBUG oslo_vmware.api [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356233, 'name': ReconfigVM_Task, 'duration_secs': 0.274759} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.470408] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 5a8cdb36-1845-4806-a6bc-22afb47886e2/5a8cdb36-1845-4806-a6bc-22afb47886e2.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1134.470710] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating instance '5a8cdb36-1845-4806-a6bc-22afb47886e2' progress to 50 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1134.474653] env[61970]: INFO nova.compute.claims [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1134.895297] env[61970]: DEBUG oslo_vmware.api [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356234, 'name': ReconfigVM_Task, 'duration_secs': 0.337528} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.896024] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Reconfigured VM instance instance-00000069 to attach disk [datastore1] volume-6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe/volume-6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1134.900442] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1de53874-0a6f-49c0-80aa-609ce58c39d3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.914937] env[61970]: DEBUG oslo_vmware.api [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1134.914937] env[61970]: value = "task-1356235" [ 1134.914937] env[61970]: _type = "Task" [ 1134.914937] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.922572] env[61970]: DEBUG oslo_vmware.api [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356235, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.982119] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f7d8f7-19d8-46b9-82dd-41f047562d5f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.003954] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb07d41f-f18d-4521-abf9-b1696560b34b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.024646] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating instance '5a8cdb36-1845-4806-a6bc-22afb47886e2' progress to 67 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1135.424897] env[61970]: DEBUG oslo_vmware.api [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356235, 'name': ReconfigVM_Task, 'duration_secs': 0.1369} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.425224] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288891', 'volume_id': '6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe', 'name': 'volume-6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'df0a2338-07e0-4148-8729-a290b88207b3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe', 'serial': '6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1135.589496] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18f9fff-9052-4fcd-a224-fa4f1d264ab5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.597144] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c8f0259-060f-4d11-887a-e26d0a3b927c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.628902] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e2286ec-ede3-4ebb-ac9a-ce1f22dd7ea7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.635824] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371da4b7-f233-4ed2-8e3a-c604efe9ea95 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.648564] env[61970]: DEBUG nova.compute.provider_tree [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1136.151868] env[61970]: DEBUG nova.scheduler.client.report [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1136.460906] env[61970]: DEBUG nova.objects.instance [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lazy-loading 'flavor' on Instance uuid df0a2338-07e0-4148-8729-a290b88207b3 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1136.657023] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.190s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.658530] env[61970]: DEBUG oslo_concurrency.lockutils [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 3.085s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.714541] env[61970]: INFO nova.network.neutron [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Updating port 2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1136.736853] env[61970]: DEBUG nova.network.neutron [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Port 696b0179-1272-4ae7-ab15-727fe4682c42 binding to destination host cpu-1 is already ACTIVE {{(pid=61970) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1136.966054] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9439eb52-c308-4897-83b5-f710875e8bd6 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "df0a2338-07e0-4148-8729-a290b88207b3" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.247s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.186112] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "df0a2338-07e0-4148-8729-a290b88207b3" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.186371] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "df0a2338-07e0-4148-8729-a290b88207b3" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.265105] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a8f45df-337b-4f58-93f9-22b39b1e7248 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.273080] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a84ac780-e4b6-4051-8c59-0dd832403225 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.302474] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bbf00e3-c669-43bb-b8ff-5867389fe5dd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.309916] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bac0a41-d9d1-44c2-9e57-3249513c1261 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.323854] env[61970]: DEBUG nova.compute.provider_tree [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1137.689659] env[61970]: INFO nova.compute.manager [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Detaching volume 6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe [ 1137.725405] env[61970]: INFO nova.virt.block_device [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Attempting to driver detach volume 6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe from mountpoint /dev/sdb [ 1137.725643] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Volume detach. Driver type: vmdk {{(pid=61970) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1137.725836] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288891', 'volume_id': '6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe', 'name': 'volume-6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'df0a2338-07e0-4148-8729-a290b88207b3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe', 'serial': '6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1137.726728] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb33849-2048-4bad-8851-238480482af5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.756199] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc697bd0-4980-494a-bcb9-7e985c031a6a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.765292] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "5a8cdb36-1845-4806-a6bc-22afb47886e2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.765543] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "5a8cdb36-1845-4806-a6bc-22afb47886e2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.765731] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "5a8cdb36-1845-4806-a6bc-22afb47886e2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.771639] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9818085c-26ec-4b86-b63a-4a25ff721536 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.792624] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-830a177d-cd7b-4974-bda6-8302e3583e24 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.807913] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] The volume has not been displaced from its original location: [datastore1] volume-6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe/volume-6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe.vmdk. No consolidation needed. {{(pid=61970) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1137.813304] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Reconfiguring VM instance instance-00000069 to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1137.813839] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24d56668-1b38-44b3-90de-b8081efee7f2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.827036] env[61970]: DEBUG nova.scheduler.client.report [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1137.835699] env[61970]: DEBUG oslo_vmware.api [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1137.835699] env[61970]: value = "task-1356237" [ 1137.835699] env[61970]: _type = "Task" [ 1137.835699] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.843864] env[61970]: DEBUG oslo_vmware.api [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356237, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.118257] env[61970]: DEBUG nova.compute.manager [req-96efc3f8-b4aa-429a-8f39-f3568c8d5612 req-585c22e0-b150-459f-a1a9-92ee5dd42c51 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Received event network-vif-plugged-2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1138.118555] env[61970]: DEBUG oslo_concurrency.lockutils [req-96efc3f8-b4aa-429a-8f39-f3568c8d5612 req-585c22e0-b150-459f-a1a9-92ee5dd42c51 service nova] Acquiring lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.118790] env[61970]: DEBUG oslo_concurrency.lockutils [req-96efc3f8-b4aa-429a-8f39-f3568c8d5612 req-585c22e0-b150-459f-a1a9-92ee5dd42c51 service nova] Lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.119187] env[61970]: DEBUG oslo_concurrency.lockutils [req-96efc3f8-b4aa-429a-8f39-f3568c8d5612 req-585c22e0-b150-459f-a1a9-92ee5dd42c51 service nova] Lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.119452] env[61970]: DEBUG nova.compute.manager [req-96efc3f8-b4aa-429a-8f39-f3568c8d5612 req-585c22e0-b150-459f-a1a9-92ee5dd42c51 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] No waiting events found dispatching network-vif-plugged-2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1138.119698] env[61970]: WARNING nova.compute.manager [req-96efc3f8-b4aa-429a-8f39-f3568c8d5612 req-585c22e0-b150-459f-a1a9-92ee5dd42c51 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Received unexpected event network-vif-plugged-2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 for instance with vm_state shelved_offloaded and task_state spawning. [ 1138.196734] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1138.197034] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquired lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.197224] env[61970]: DEBUG nova.network.neutron [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1138.345904] env[61970]: DEBUG oslo_vmware.api [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356237, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.817603] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1138.817920] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.817965] env[61970]: DEBUG nova.network.neutron [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1138.835689] env[61970]: DEBUG oslo_concurrency.lockutils [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.177s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.847704] env[61970]: DEBUG oslo_vmware.api [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356237, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.958824] env[61970]: DEBUG nova.network.neutron [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Updating instance_info_cache with network_info: [{"id": "2fa02dfe-2cb0-4a72-bc30-4f991339dcb9", "address": "fa:16:3e:df:19:94", "network": {"id": "70081995-3d6c-466e-b811-1c8e532e552f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-641102750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1d4724cbb0841859369af01f2c03db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aaf1b231-3660-4453-b4f3-44d825b9a5dd", "external-id": "nsx-vlan-transportzone-6", "segmentation_id": 6, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fa02dfe-2c", "ovs_interfaceid": "2fa02dfe-2cb0-4a72-bc30-4f991339dcb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.351407] env[61970]: DEBUG oslo_vmware.api [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356237, 'name': ReconfigVM_Task, 'duration_secs': 1.241892} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.353752] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Reconfigured VM instance instance-00000069 to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1139.360566] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0079a38a-c596-4d62-a833-09b7c5475793 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.374878] env[61970]: DEBUG oslo_vmware.api [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1139.374878] env[61970]: value = "task-1356239" [ 1139.374878] env[61970]: _type = "Task" [ 1139.374878] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.382335] env[61970]: DEBUG oslo_vmware.api [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356239, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.402866] env[61970]: INFO nova.scheduler.client.report [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleted allocation for migration a273a020-6905-4994-806d-f9aa5f5c9d19 [ 1139.462091] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Releasing lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1139.489930] env[61970]: DEBUG nova.virt.hardware [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='3bc0ad94d771d758df83f8dff70f6cd9',container_format='bare',created_at=2024-10-24T13:47:44Z,direct_url=,disk_format='vmdk',id=e7e95577-cfb6-4cb0-9088-e43964fd7b75,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-2003574578-shelved',owner='e1d4724cbb0841859369af01f2c03db1',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2024-10-24T13:47:56Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1139.490212] env[61970]: DEBUG nova.virt.hardware [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1139.490376] env[61970]: DEBUG nova.virt.hardware [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1139.490563] env[61970]: DEBUG nova.virt.hardware [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1139.490715] env[61970]: DEBUG nova.virt.hardware [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1139.490894] env[61970]: DEBUG nova.virt.hardware [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1139.491130] env[61970]: DEBUG nova.virt.hardware [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1139.491301] env[61970]: DEBUG nova.virt.hardware [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1139.491472] env[61970]: DEBUG nova.virt.hardware [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1139.491637] env[61970]: DEBUG nova.virt.hardware [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1139.491809] env[61970]: DEBUG nova.virt.hardware [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1139.492699] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9fdf1b4-d757-4a21-9cc2-c3ed033faf64 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.502394] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-347fc07f-6b1b-4764-8997-65bd81b57c3a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.516047] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:df:19:94', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aaf1b231-3660-4453-b4f3-44d825b9a5dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2fa02dfe-2cb0-4a72-bc30-4f991339dcb9', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1139.522776] env[61970]: DEBUG oslo.service.loopingcall [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1139.523015] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1139.523236] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7e0004f2-8c48-432f-9559-08420929acc8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.542811] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1139.542811] env[61970]: value = "task-1356240" [ 1139.542811] env[61970]: _type = "Task" [ 1139.542811] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.549790] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356240, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.569911] env[61970]: DEBUG nova.network.neutron [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating instance_info_cache with network_info: [{"id": "696b0179-1272-4ae7-ab15-727fe4682c42", "address": "fa:16:3e:9e:3f:74", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap696b0179-12", "ovs_interfaceid": "696b0179-1272-4ae7-ab15-727fe4682c42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.885781] env[61970]: DEBUG oslo_vmware.api [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356239, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.911101] env[61970]: DEBUG oslo_concurrency.lockutils [None req-40f52fe9-a198-4c6e-b1be-0a1e9c3806e4 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.755s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.049921] env[61970]: INFO nova.compute.manager [None req-48a130be-6247-4f7d-854a-849d2a1babba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Get console output [ 1140.050289] env[61970]: WARNING nova.virt.vmwareapi.driver [None req-48a130be-6247-4f7d-854a-849d2a1babba tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] The console log is missing. Check your VSPC configuration [ 1140.055859] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356240, 'name': CreateVM_Task, 'duration_secs': 0.3964} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.056229] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1140.056807] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7e95577-cfb6-4cb0-9088-e43964fd7b75" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.056968] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7e95577-cfb6-4cb0-9088-e43964fd7b75" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.057349] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7e95577-cfb6-4cb0-9088-e43964fd7b75" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1140.057591] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-efcd145b-407d-4733-9aee-98c19c40ee99 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.062511] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1140.062511] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5294554c-8e57-1889-526c-65b3341e8a87" [ 1140.062511] env[61970]: _type = "Task" [ 1140.062511] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.070295] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5294554c-8e57-1889-526c-65b3341e8a87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.072797] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.142388] env[61970]: DEBUG nova.compute.manager [req-2e342db7-6780-44db-893c-e04da64cb682 req-c8dbebdb-1663-4816-b2bc-3d19204b7c54 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Received event network-changed-2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1140.142589] env[61970]: DEBUG nova.compute.manager [req-2e342db7-6780-44db-893c-e04da64cb682 req-c8dbebdb-1663-4816-b2bc-3d19204b7c54 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Refreshing instance network info cache due to event network-changed-2fa02dfe-2cb0-4a72-bc30-4f991339dcb9. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1140.142818] env[61970]: DEBUG oslo_concurrency.lockutils [req-2e342db7-6780-44db-893c-e04da64cb682 req-c8dbebdb-1663-4816-b2bc-3d19204b7c54 service nova] Acquiring lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.142988] env[61970]: DEBUG oslo_concurrency.lockutils [req-2e342db7-6780-44db-893c-e04da64cb682 req-c8dbebdb-1663-4816-b2bc-3d19204b7c54 service nova] Acquired lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.143176] env[61970]: DEBUG nova.network.neutron [req-2e342db7-6780-44db-893c-e04da64cb682 req-c8dbebdb-1663-4816-b2bc-3d19204b7c54 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Refreshing network info cache for port 2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1140.385289] env[61970]: DEBUG oslo_vmware.api [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356239, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.572889] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7e95577-cfb6-4cb0-9088-e43964fd7b75" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.573161] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Processing image e7e95577-cfb6-4cb0-9088-e43964fd7b75 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1140.573402] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7e95577-cfb6-4cb0-9088-e43964fd7b75/e7e95577-cfb6-4cb0-9088-e43964fd7b75.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.573553] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7e95577-cfb6-4cb0-9088-e43964fd7b75/e7e95577-cfb6-4cb0-9088-e43964fd7b75.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.573735] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1140.573984] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-21157397-fc6a-4f67-97c7-3bfbb0db1121 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.581058] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75a505fc-0b62-4c62-bfe5-66fe1719438f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.584204] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1140.584381] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1140.585315] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33174dec-9e8a-4c64-bfa8-e1518ca670b7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.589456] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc9bb4c8-737f-4a56-9b16-4ab5b224ba1a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.592469] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1140.592469] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f1360d-803b-ab86-918b-bdc940958f06" [ 1140.592469] env[61970]: _type = "Task" [ 1140.592469] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.600206] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52f1360d-803b-ab86-918b-bdc940958f06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.865184] env[61970]: DEBUG nova.network.neutron [req-2e342db7-6780-44db-893c-e04da64cb682 req-c8dbebdb-1663-4816-b2bc-3d19204b7c54 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Updated VIF entry in instance network info cache for port 2fa02dfe-2cb0-4a72-bc30-4f991339dcb9. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1140.865566] env[61970]: DEBUG nova.network.neutron [req-2e342db7-6780-44db-893c-e04da64cb682 req-c8dbebdb-1663-4816-b2bc-3d19204b7c54 service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Updating instance_info_cache with network_info: [{"id": "2fa02dfe-2cb0-4a72-bc30-4f991339dcb9", "address": "fa:16:3e:df:19:94", "network": {"id": "70081995-3d6c-466e-b811-1c8e532e552f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-641102750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1d4724cbb0841859369af01f2c03db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aaf1b231-3660-4453-b4f3-44d825b9a5dd", "external-id": "nsx-vlan-transportzone-6", "segmentation_id": 6, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fa02dfe-2c", "ovs_interfaceid": "2fa02dfe-2cb0-4a72-bc30-4f991339dcb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.885379] env[61970]: DEBUG oslo_vmware.api [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356239, 'name': ReconfigVM_Task, 'duration_secs': 1.116656} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.885612] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288891', 'volume_id': '6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe', 'name': 'volume-6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'df0a2338-07e0-4148-8729-a290b88207b3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe', 'serial': '6c4b9f9c-d11d-49e3-9de6-605b7b0c77fe'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1141.102668] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Preparing fetch location {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1141.102939] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Fetch image to [datastore1] OSTACK_IMG_ebf67bbf-9b17-4111-b9e8-ed5638e82866/OSTACK_IMG_ebf67bbf-9b17-4111-b9e8-ed5638e82866.vmdk {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1141.103117] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Downloading stream optimized image e7e95577-cfb6-4cb0-9088-e43964fd7b75 to [datastore1] OSTACK_IMG_ebf67bbf-9b17-4111-b9e8-ed5638e82866/OSTACK_IMG_ebf67bbf-9b17-4111-b9e8-ed5638e82866.vmdk on the data store datastore1 as vApp {{(pid=61970) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1141.103294] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Downloading image file data e7e95577-cfb6-4cb0-9088-e43964fd7b75 to the ESX as VM named 'OSTACK_IMG_ebf67bbf-9b17-4111-b9e8-ed5638e82866' {{(pid=61970) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1141.169072] env[61970]: DEBUG oslo_vmware.rw_handles [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1141.169072] env[61970]: value = "resgroup-9" [ 1141.169072] env[61970]: _type = "ResourcePool" [ 1141.169072] env[61970]: }. {{(pid=61970) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1141.169353] env[61970]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-0ec4f488-89cb-48bb-a386-8bebb2e094bd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.190171] env[61970]: DEBUG oslo_vmware.rw_handles [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lease: (returnval){ [ 1141.190171] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523cb1e5-48e1-5319-dab9-e949cd8ce435" [ 1141.190171] env[61970]: _type = "HttpNfcLease" [ 1141.190171] env[61970]: } obtained for vApp import into resource pool (val){ [ 1141.190171] env[61970]: value = "resgroup-9" [ 1141.190171] env[61970]: _type = "ResourcePool" [ 1141.190171] env[61970]: }. {{(pid=61970) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1141.190467] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the lease: (returnval){ [ 1141.190467] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523cb1e5-48e1-5319-dab9-e949cd8ce435" [ 1141.190467] env[61970]: _type = "HttpNfcLease" [ 1141.190467] env[61970]: } to be ready. {{(pid=61970) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1141.196390] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1141.196390] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523cb1e5-48e1-5319-dab9-e949cd8ce435" [ 1141.196390] env[61970]: _type = "HttpNfcLease" [ 1141.196390] env[61970]: } is initializing. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1141.367996] env[61970]: DEBUG oslo_concurrency.lockutils [req-2e342db7-6780-44db-893c-e04da64cb682 req-c8dbebdb-1663-4816-b2bc-3d19204b7c54 service nova] Releasing lock "refresh_cache-3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.425605] env[61970]: DEBUG nova.objects.instance [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lazy-loading 'flavor' on Instance uuid df0a2338-07e0-4148-8729-a290b88207b3 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1141.685792] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322f375c-c2d4-41b5-9b51-f1438fc62ac0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.711589] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83d5d0e-3b9b-4fcc-863b-bbfa5ee19d06 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.717972] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1141.717972] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523cb1e5-48e1-5319-dab9-e949cd8ce435" [ 1141.717972] env[61970]: _type = "HttpNfcLease" [ 1141.717972] env[61970]: } is ready. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1141.720291] env[61970]: DEBUG oslo_vmware.rw_handles [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1141.720291] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523cb1e5-48e1-5319-dab9-e949cd8ce435" [ 1141.720291] env[61970]: _type = "HttpNfcLease" [ 1141.720291] env[61970]: }. {{(pid=61970) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1141.720652] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating instance '5a8cdb36-1845-4806-a6bc-22afb47886e2' progress to 83 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1141.724636] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a051b03-333b-490c-8136-a6e30f47f147 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.731978] env[61970]: DEBUG oslo_vmware.rw_handles [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527251c3-6de5-7794-c895-9aef6eee6547/disk-0.vmdk from lease info. {{(pid=61970) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1141.732202] env[61970]: DEBUG oslo_vmware.rw_handles [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527251c3-6de5-7794-c895-9aef6eee6547/disk-0.vmdk. {{(pid=61970) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1141.800882] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0a6636ef-1434-40c2-81c7-0c4db0c1aa69 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.233027] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1142.233027] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d15e8ea-1a2a-4814-8e50-a92e71d3b387 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.239744] env[61970]: DEBUG oslo_vmware.api [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1142.239744] env[61970]: value = "task-1356242" [ 1142.239744] env[61970]: _type = "Task" [ 1142.239744] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.250407] env[61970]: DEBUG oslo_vmware.api [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356242, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.434355] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b5000dd0-755b-456b-80ab-b12e21bc9165 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "df0a2338-07e0-4148-8729-a290b88207b3" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 5.248s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.750297] env[61970]: DEBUG oslo_vmware.api [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356242, 'name': PowerOnVM_Task, 'duration_secs': 0.397943} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.750604] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1142.750812] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6deeecbc-994e-4296-ba00-ea337ed8c15d tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating instance '5a8cdb36-1845-4806-a6bc-22afb47886e2' progress to 100 {{(pid=61970) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1142.916784] env[61970]: DEBUG oslo_vmware.rw_handles [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Completed reading data from the image iterator. {{(pid=61970) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1142.917116] env[61970]: DEBUG oslo_vmware.rw_handles [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527251c3-6de5-7794-c895-9aef6eee6547/disk-0.vmdk. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1142.918242] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beff6338-5f3d-42d6-8663-d41dc4740036 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.924964] env[61970]: DEBUG oslo_vmware.rw_handles [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527251c3-6de5-7794-c895-9aef6eee6547/disk-0.vmdk is in state: ready. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1142.925069] env[61970]: DEBUG oslo_vmware.rw_handles [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527251c3-6de5-7794-c895-9aef6eee6547/disk-0.vmdk. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1142.925321] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-67744b78-7e40-4ed7-aca4-5471f0625933 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.113186] env[61970]: DEBUG oslo_vmware.rw_handles [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527251c3-6de5-7794-c895-9aef6eee6547/disk-0.vmdk. {{(pid=61970) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1143.113425] env[61970]: INFO nova.virt.vmwareapi.images [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Downloaded image file data e7e95577-cfb6-4cb0-9088-e43964fd7b75 [ 1143.114353] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559c337f-7bd0-45bc-9130-88d1918ef665 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.130063] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0efe820d-16df-4928-a827-b91328fdc35c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.172595] env[61970]: INFO nova.virt.vmwareapi.images [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] The imported VM was unregistered [ 1143.174929] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Caching image {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1143.175180] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Creating directory with path [datastore1] devstack-image-cache_base/e7e95577-cfb6-4cb0-9088-e43964fd7b75 {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1143.175452] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff10f637-5dd7-4d3d-b628-21b6e3c8fb8b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.200537] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Created directory with path [datastore1] devstack-image-cache_base/e7e95577-cfb6-4cb0-9088-e43964fd7b75 {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1143.200703] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_ebf67bbf-9b17-4111-b9e8-ed5638e82866/OSTACK_IMG_ebf67bbf-9b17-4111-b9e8-ed5638e82866.vmdk to [datastore1] devstack-image-cache_base/e7e95577-cfb6-4cb0-9088-e43964fd7b75/e7e95577-cfb6-4cb0-9088-e43964fd7b75.vmdk. {{(pid=61970) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1143.200985] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-fc12f0ee-f426-4cfa-a83b-e025f80768b9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.209307] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1143.209307] env[61970]: value = "task-1356244" [ 1143.209307] env[61970]: _type = "Task" [ 1143.209307] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.217725] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356244, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.511504] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "df0a2338-07e0-4148-8729-a290b88207b3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.511958] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "df0a2338-07e0-4148-8729-a290b88207b3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.511958] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "df0a2338-07e0-4148-8729-a290b88207b3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.512181] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "df0a2338-07e0-4148-8729-a290b88207b3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.512332] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "df0a2338-07e0-4148-8729-a290b88207b3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.514611] env[61970]: INFO nova.compute.manager [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Terminating instance [ 1143.516396] env[61970]: DEBUG nova.compute.manager [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1143.516596] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1143.517532] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbbffd19-0628-4ef9-b715-b9a897e4b539 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.525010] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1143.525260] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-abdddeb3-8069-437c-8fd0-ac351ca266d3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.531154] env[61970]: DEBUG oslo_vmware.api [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1143.531154] env[61970]: value = "task-1356245" [ 1143.531154] env[61970]: _type = "Task" [ 1143.531154] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.538922] env[61970]: DEBUG oslo_vmware.api [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356245, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.718676] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356244, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.041252] env[61970]: DEBUG oslo_vmware.api [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356245, 'name': PowerOffVM_Task, 'duration_secs': 0.425768} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.041544] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1144.041718] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1144.042411] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9d4a8a74-ee49-4a9e-aff4-72557e21c6c1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.146033] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1144.146392] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1144.146603] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Deleting the datastore file [datastore2] df0a2338-07e0-4148-8729-a290b88207b3 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1144.147042] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cad8ff25-55ca-4a20-8e02-5ac10051b9a1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.153647] env[61970]: DEBUG oslo_vmware.api [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for the task: (returnval){ [ 1144.153647] env[61970]: value = "task-1356247" [ 1144.153647] env[61970]: _type = "Task" [ 1144.153647] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.162182] env[61970]: DEBUG oslo_vmware.api [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356247, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.219113] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356244, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.669018] env[61970]: DEBUG oslo_vmware.api [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Task: {'id': task-1356247, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209417} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.669435] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1144.669819] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1144.670015] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1144.670324] env[61970]: INFO nova.compute.manager [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1144.670686] env[61970]: DEBUG oslo.service.loopingcall [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1144.671224] env[61970]: DEBUG nova.compute.manager [-] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1144.671386] env[61970]: DEBUG nova.network.neutron [-] [instance: df0a2338-07e0-4148-8729-a290b88207b3] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1144.720129] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356244, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.789147] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1145.199995] env[61970]: DEBUG nova.network.neutron [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Port 696b0179-1272-4ae7-ab15-727fe4682c42 binding to destination host cpu-1 is already ACTIVE {{(pid=61970) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1145.200387] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1145.201093] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.201352] env[61970]: DEBUG nova.network.neutron [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1145.221654] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356244, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.255857] env[61970]: DEBUG nova.compute.manager [req-1ee79d61-7c99-4c44-a02e-8f3a3aec94ee req-20b9a5fe-15d8-4ec4-a4f2-46ec52989aab service nova] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Received event network-vif-deleted-8aabb66f-904e-4c84-b459-c3c3ac76d349 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1145.256129] env[61970]: INFO nova.compute.manager [req-1ee79d61-7c99-4c44-a02e-8f3a3aec94ee req-20b9a5fe-15d8-4ec4-a4f2-46ec52989aab service nova] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Neutron deleted interface 8aabb66f-904e-4c84-b459-c3c3ac76d349; detaching it from the instance and deleting it from the info cache [ 1145.256315] env[61970]: DEBUG nova.network.neutron [req-1ee79d61-7c99-4c44-a02e-8f3a3aec94ee req-20b9a5fe-15d8-4ec4-a4f2-46ec52989aab service nova] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.259391] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1145.616297] env[61970]: DEBUG nova.network.neutron [-] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.724012] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356244, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.760154] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1b135df4-f1d0-48f4-95f3-64f1b8b2d91a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.769888] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6430bf92-a0b0-4d12-b340-01f3eefd7e15 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.800116] env[61970]: DEBUG nova.compute.manager [req-1ee79d61-7c99-4c44-a02e-8f3a3aec94ee req-20b9a5fe-15d8-4ec4-a4f2-46ec52989aab service nova] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Detach interface failed, port_id=8aabb66f-904e-4c84-b459-c3c3ac76d349, reason: Instance df0a2338-07e0-4148-8729-a290b88207b3 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1145.987111] env[61970]: DEBUG nova.network.neutron [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating instance_info_cache with network_info: [{"id": "696b0179-1272-4ae7-ab15-727fe4682c42", "address": "fa:16:3e:9e:3f:74", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap696b0179-12", "ovs_interfaceid": "696b0179-1272-4ae7-ab15-727fe4682c42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.118752] env[61970]: INFO nova.compute.manager [-] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Took 1.45 seconds to deallocate network for instance. [ 1146.223962] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356244, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.489403] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.625061] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.625339] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.625569] env[61970]: DEBUG nova.objects.instance [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lazy-loading 'resources' on Instance uuid df0a2338-07e0-4148-8729-a290b88207b3 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1146.721450] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356244, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.993178] env[61970]: DEBUG nova.compute.manager [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61970) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1147.218451] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba5ee1b9-32a8-4f28-a201-10f00a69a340 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.223820] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356244, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.733274} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.224374] env[61970]: INFO nova.virt.vmwareapi.ds_util [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_ebf67bbf-9b17-4111-b9e8-ed5638e82866/OSTACK_IMG_ebf67bbf-9b17-4111-b9e8-ed5638e82866.vmdk to [datastore1] devstack-image-cache_base/e7e95577-cfb6-4cb0-9088-e43964fd7b75/e7e95577-cfb6-4cb0-9088-e43964fd7b75.vmdk. [ 1147.224570] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Cleaning up location [datastore1] OSTACK_IMG_ebf67bbf-9b17-4111-b9e8-ed5638e82866 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1147.224738] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_ebf67bbf-9b17-4111-b9e8-ed5638e82866 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1147.224966] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-97d4cbd1-18ef-402c-80b2-8b8b74d74b24 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.228854] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7afedb8-6906-4722-afdf-f4c8e7e70e90 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.232476] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1147.232476] env[61970]: value = "task-1356248" [ 1147.232476] env[61970]: _type = "Task" [ 1147.232476] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.258767] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c4aad4-5496-41d1-9ff3-b8b9ea6575e0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.263487] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1147.263643] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61970) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1147.263855] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356248, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.264054] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1147.269377] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eb1740f-5c77-4a75-9df8-71ef4e885628 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.282514] env[61970]: DEBUG nova.compute.provider_tree [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1147.745705] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356248, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.109889} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.745994] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1147.746188] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7e95577-cfb6-4cb0-9088-e43964fd7b75/e7e95577-cfb6-4cb0-9088-e43964fd7b75.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.746439] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7e95577-cfb6-4cb0-9088-e43964fd7b75/e7e95577-cfb6-4cb0-9088-e43964fd7b75.vmdk to [datastore1] 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6/3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1147.746701] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1216b769-cdad-409b-83bf-7084642cd218 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.752457] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1147.752457] env[61970]: value = "task-1356249" [ 1147.752457] env[61970]: _type = "Task" [ 1147.752457] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.759891] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356249, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.766878] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.802183] env[61970]: ERROR nova.scheduler.client.report [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] [req-ad28985f-8697-41eb-85c3-bd56a28404ab] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fc0f3e4e-5ef5-45e9-8335-cc68743405aa. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ad28985f-8697-41eb-85c3-bd56a28404ab"}]} [ 1147.822498] env[61970]: DEBUG nova.scheduler.client.report [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Refreshing inventories for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1147.835718] env[61970]: DEBUG nova.scheduler.client.report [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Updating ProviderTree inventory for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1147.835974] env[61970]: DEBUG nova.compute.provider_tree [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1147.846934] env[61970]: DEBUG nova.scheduler.client.report [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Refreshing aggregate associations for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa, aggregates: None {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1147.863858] env[61970]: DEBUG nova.scheduler.client.report [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Refreshing trait associations for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1147.952590] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ede95065-d0e4-447f-aada-b5284bf5779b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.963066] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-675d6d64-7c05-4b3c-b494-c4ab65f7e674 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.998064] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4097831c-ea29-415d-aac2-bd2792451670 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.006454] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262ae5aa-3a18-4e8a-8994-61b43b3ddeab {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.022184] env[61970]: DEBUG nova.compute.provider_tree [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1148.084305] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.263214] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356249, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.553414] env[61970]: DEBUG nova.scheduler.client.report [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Updated inventory for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with generation 141 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1148.553738] env[61970]: DEBUG nova.compute.provider_tree [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Updating resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa generation from 141 to 142 during operation: update_inventory {{(pid=61970) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1148.553952] env[61970]: DEBUG nova.compute.provider_tree [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1148.764386] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356249, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.059684] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.434s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.063361] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.296s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.063633] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.063824] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61970) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1149.064184] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.980s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.066384] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d761f9-c148-42e4-bf97-ed2ac918a6f7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.075872] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded606ab-6f1d-4b91-beb2-ff0c450fdcbe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.083761] env[61970]: INFO nova.scheduler.client.report [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Deleted allocations for instance df0a2338-07e0-4148-8729-a290b88207b3 [ 1149.096029] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b41b22-66ea-4b2f-97af-2642388d00e9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.104984] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda4959a-708e-4825-8472-17e4a0888f3f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.137240] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180479MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61970) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1149.137480] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.264949] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356249, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.570242] env[61970]: DEBUG nova.objects.instance [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lazy-loading 'migration_context' on Instance uuid 5a8cdb36-1845-4806-a6bc-22afb47886e2 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1149.603581] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e2226497-aa9a-43bd-92c1-f27dfec6f241 tempest-AttachVolumeNegativeTest-961921540 tempest-AttachVolumeNegativeTest-961921540-project-member] Lock "df0a2338-07e0-4148-8729-a290b88207b3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.092s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.765972] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356249, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.158580] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6d1ae35-8d95-4fab-a693-14e65c045ef7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.168145] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b6413c7-30f3-4e2e-a2b0-8ef291ff9671 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.197987] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cdc755d-8fa4-4454-8f82-2dce50fafdfa {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.206178] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3424dff-5c37-461e-95b3-335fc400c3aa {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.220164] env[61970]: DEBUG nova.compute.provider_tree [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1150.264300] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356249, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.31481} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.264536] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7e95577-cfb6-4cb0-9088-e43964fd7b75/e7e95577-cfb6-4cb0-9088-e43964fd7b75.vmdk to [datastore1] 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6/3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1150.265305] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db9d64f9-5717-49ee-8290-912084e0668e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.286435] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6/3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6.vmdk or device None with type streamOptimized {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1150.286714] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0640fac-1a4e-440f-8bef-8da401d0da2e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.307361] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1150.307361] env[61970]: value = "task-1356251" [ 1150.307361] env[61970]: _type = "Task" [ 1150.307361] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.317598] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356251, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.723983] env[61970]: DEBUG nova.scheduler.client.report [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1150.818754] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356251, 'name': ReconfigVM_Task, 'duration_secs': 0.284051} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.818998] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6/3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6.vmdk or device None with type streamOptimized {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1150.819592] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0c8cf29f-2797-4e2c-b58a-8857cb18e1e5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.825039] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1150.825039] env[61970]: value = "task-1356252" [ 1150.825039] env[61970]: _type = "Task" [ 1150.825039] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.831944] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356252, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.334597] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356252, 'name': Rename_Task, 'duration_secs': 0.167147} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.334943] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1151.335050] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-89c01667-71f0-4129-9134-67e81064ee77 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.341820] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1151.341820] env[61970]: value = "task-1356253" [ 1151.341820] env[61970]: _type = "Task" [ 1151.341820] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.348803] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356253, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.735230] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.671s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.740932] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.603s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.851075] env[61970]: DEBUG oslo_vmware.api [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356253, 'name': PowerOnVM_Task, 'duration_secs': 0.435902} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.851342] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1151.944665] env[61970]: DEBUG nova.compute.manager [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1151.945586] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b68e13-f022-4e94-92bc-b3449e00563c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.465774] env[61970]: DEBUG oslo_concurrency.lockutils [None req-a332bf7d-fc58-4a17-8e20-6c655de4b445 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 20.983s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.777711] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 0304640a-6c46-4f57-a8b8-5058da2e3de4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1152.777876] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 183133c6-7e4e-4dd4-b8e0-cb4f8338264c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1152.778010] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1153.278037] env[61970]: INFO nova.compute.manager [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Swapping old allocation on dict_keys(['fc0f3e4e-5ef5-45e9-8335-cc68743405aa']) held by migration 835beeae-86b7-4c24-85fe-065e481c60a0 for instance [ 1153.280191] env[61970]: INFO nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 835beeae-86b7-4c24-85fe-065e481c60a0 has allocations against this compute host but is not found in the database. [ 1153.280393] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 5a8cdb36-1845-4806-a6bc-22afb47886e2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1153.280617] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1153.280846] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1153.281019] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1600MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1153.307912] env[61970]: DEBUG nova.scheduler.client.report [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Overwriting current allocation {'allocations': {'fc0f3e4e-5ef5-45e9-8335-cc68743405aa': {'resources': {'VCPU': 1, 'MEMORY_MB': 256, 'DISK_GB': 1}, 'generation': 142}}, 'project_id': '09033ee688384ca287b4fdad6e67cb1f', 'user_id': 'd2b8322c5de6483aaf032781d28e950d', 'consumer_generation': 1} on consumer 5a8cdb36-1845-4806-a6bc-22afb47886e2 {{(pid=61970) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1153.351260] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07df92a-ff5d-4099-8230-542e77d24417 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.358434] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e106ffb-2803-4be0-83a7-0575c557103f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.390711] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d687e6d-5371-411b-87ce-f5741329534d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.393817] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.393987] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.394180] env[61970]: DEBUG nova.network.neutron [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1153.400084] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74165908-ff0e-49d4-be73-2b5c74e88e76 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.413298] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1153.915915] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1154.098982] env[61970]: DEBUG nova.network.neutron [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating instance_info_cache with network_info: [{"id": "696b0179-1272-4ae7-ab15-727fe4682c42", "address": "fa:16:3e:9e:3f:74", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap696b0179-12", "ovs_interfaceid": "696b0179-1272-4ae7-ab15-727fe4682c42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.421341] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1154.421563] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.681s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.601683] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "refresh_cache-5a8cdb36-1845-4806-a6bc-22afb47886e2" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.602795] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-defa251d-c9b7-41a5-b08b-4e99aafa2d61 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.610233] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43fed827-ad1e-4a6f-a94b-aec5cbb2e74b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.413344] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1155.413825] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1155.413825] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Starting heal instance info cache {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1155.413944] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Rebuilding the list of instances to heal {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1155.688210] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1155.688814] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00de2524-4e4b-4601-8840-78cba48ac0a9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.698836] env[61970]: DEBUG oslo_vmware.api [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1155.698836] env[61970]: value = "task-1356255" [ 1155.698836] env[61970]: _type = "Task" [ 1155.698836] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.706927] env[61970]: DEBUG oslo_vmware.api [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356255, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.960801] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "refresh_cache-0304640a-6c46-4f57-a8b8-5058da2e3de4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1155.961036] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquired lock "refresh_cache-0304640a-6c46-4f57-a8b8-5058da2e3de4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.961215] env[61970]: DEBUG nova.network.neutron [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Forcefully refreshing network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1155.961387] env[61970]: DEBUG nova.objects.instance [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lazy-loading 'info_cache' on Instance uuid 0304640a-6c46-4f57-a8b8-5058da2e3de4 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1156.208973] env[61970]: DEBUG oslo_vmware.api [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356255, 'name': PowerOffVM_Task, 'duration_secs': 0.297455} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.209265] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1156.209998] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1156.210255] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1156.210376] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1156.210577] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1156.210770] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1156.210919] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1156.211143] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1156.211313] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1156.211483] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1156.211647] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1156.211819] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1156.216598] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2cbe948-bf1b-47e8-949a-def08ab1c93f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.232621] env[61970]: DEBUG oslo_vmware.api [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1156.232621] env[61970]: value = "task-1356256" [ 1156.232621] env[61970]: _type = "Task" [ 1156.232621] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.240198] env[61970]: DEBUG oslo_vmware.api [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356256, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.742580] env[61970]: DEBUG oslo_vmware.api [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356256, 'name': ReconfigVM_Task, 'duration_secs': 0.314991} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.743351] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8741239-5db9-4df4-814f-ac66001bbc03 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.763428] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1156.763668] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1156.763865] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1156.764368] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1156.764368] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1156.764368] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1156.764562] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1156.764724] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1156.764895] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1156.765074] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1156.765256] env[61970]: DEBUG nova.virt.hardware [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1156.765997] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7862bfc6-f690-4317-b308-3f334d9af57c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.771066] env[61970]: DEBUG oslo_vmware.api [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1156.771066] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d1adad-4512-40bc-fed8-6c4a530ccaf2" [ 1156.771066] env[61970]: _type = "Task" [ 1156.771066] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.778305] env[61970]: DEBUG oslo_vmware.api [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d1adad-4512-40bc-fed8-6c4a530ccaf2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.280839] env[61970]: DEBUG oslo_vmware.api [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d1adad-4512-40bc-fed8-6c4a530ccaf2, 'name': SearchDatastore_Task, 'duration_secs': 0.011149} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.286043] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Reconfiguring VM instance instance-00000068 to detach disk 2000 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1157.286310] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5cd76f1-f97c-405d-90ca-82034fd23f8d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.303928] env[61970]: DEBUG oslo_vmware.api [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1157.303928] env[61970]: value = "task-1356258" [ 1157.303928] env[61970]: _type = "Task" [ 1157.303928] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.316096] env[61970]: DEBUG oslo_vmware.api [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356258, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.663387] env[61970]: DEBUG nova.network.neutron [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Updating instance_info_cache with network_info: [{"id": "2545e2b0-a432-44ae-a287-0b68ceab7031", "address": "fa:16:3e:a6:77:ea", "network": {"id": "cc469189-e44e-47e1-97e6-72c169c85925", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-383668965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a078b8af9c469eb279be4da7459166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2545e2b0-a4", "ovs_interfaceid": "2545e2b0-a432-44ae-a287-0b68ceab7031", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.814855] env[61970]: DEBUG oslo_vmware.api [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356258, 'name': ReconfigVM_Task, 'duration_secs': 0.195856} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.815249] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Reconfigured VM instance instance-00000068 to detach disk 2000 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1157.815822] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8efe0666-3d9a-4cce-8647-322c181096c4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.840032] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 5a8cdb36-1845-4806-a6bc-22afb47886e2/5a8cdb36-1845-4806-a6bc-22afb47886e2.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1157.840330] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3dd413a-efad-44ee-93a9-4f746b934fc2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.857988] env[61970]: DEBUG oslo_vmware.api [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1157.857988] env[61970]: value = "task-1356259" [ 1157.857988] env[61970]: _type = "Task" [ 1157.857988] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.865746] env[61970]: DEBUG oslo_vmware.api [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356259, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.166353] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Releasing lock "refresh_cache-0304640a-6c46-4f57-a8b8-5058da2e3de4" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.166589] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Updated the network info_cache for instance {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1158.166832] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1158.167013] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1158.167180] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1158.367962] env[61970]: DEBUG oslo_vmware.api [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356259, 'name': ReconfigVM_Task, 'duration_secs': 0.251072} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.368241] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 5a8cdb36-1845-4806-a6bc-22afb47886e2/5a8cdb36-1845-4806-a6bc-22afb47886e2.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1158.368996] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c8f628-ed4c-40c8-8ee3-2ea8deb9b306 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.389137] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aeb81df-5126-4f62-875a-13de4ba0e866 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.408994] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87cb123a-0a5d-4ea9-8d96-87d4ee328d64 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.428415] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4bd6f1e-618f-40ad-ada6-1c68bcd08757 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.434707] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1158.434941] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0335f64c-18c7-458b-9381-b1a742195768 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.440795] env[61970]: DEBUG oslo_vmware.api [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1158.440795] env[61970]: value = "task-1356260" [ 1158.440795] env[61970]: _type = "Task" [ 1158.440795] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.447631] env[61970]: DEBUG oslo_vmware.api [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356260, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.950564] env[61970]: DEBUG oslo_vmware.api [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356260, 'name': PowerOnVM_Task, 'duration_secs': 0.410169} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.950981] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1159.997074] env[61970]: INFO nova.compute.manager [None req-e356bf48-9bde-4a3c-aca4-bdffe9b47995 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating instance to original state: 'active' [ 1161.302115] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "5a8cdb36-1845-4806-a6bc-22afb47886e2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.302115] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "5a8cdb36-1845-4806-a6bc-22afb47886e2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.302483] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "5a8cdb36-1845-4806-a6bc-22afb47886e2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.302891] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "5a8cdb36-1845-4806-a6bc-22afb47886e2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.303128] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "5a8cdb36-1845-4806-a6bc-22afb47886e2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.305271] env[61970]: INFO nova.compute.manager [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Terminating instance [ 1161.307183] env[61970]: DEBUG nova.compute.manager [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1161.307442] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1161.307834] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe670955-43cc-455a-b365-0523273b4bd2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.315562] env[61970]: DEBUG oslo_vmware.api [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1161.315562] env[61970]: value = "task-1356261" [ 1161.315562] env[61970]: _type = "Task" [ 1161.315562] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.326816] env[61970]: DEBUG oslo_vmware.api [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356261, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.824737] env[61970]: DEBUG oslo_vmware.api [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356261, 'name': PowerOffVM_Task, 'duration_secs': 0.223956} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.824997] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1161.825213] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Volume detach. Driver type: vmdk {{(pid=61970) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1161.825409] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288890', 'volume_id': '9cb12dfc-bdee-43a7-b4ba-16190814effd', 'name': 'volume-9cb12dfc-bdee-43a7-b4ba-16190814effd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '5a8cdb36-1845-4806-a6bc-22afb47886e2', 'attached_at': '2024-10-24T13:48:27.000000', 'detached_at': '', 'volume_id': '9cb12dfc-bdee-43a7-b4ba-16190814effd', 'serial': '9cb12dfc-bdee-43a7-b4ba-16190814effd'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1161.826181] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-917ddfc1-3404-41cd-ac3f-50e4233ad1c9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.846558] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b098ca50-d329-4bba-be75-68ff5a4aac85 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.852988] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99601fe-fe8d-41bc-a465-9dce85957229 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.873383] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26f1636-a22f-44f8-8f50-debb7b06b7fa {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.887095] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] The volume has not been displaced from its original location: [datastore1] volume-9cb12dfc-bdee-43a7-b4ba-16190814effd/volume-9cb12dfc-bdee-43a7-b4ba-16190814effd.vmdk. No consolidation needed. {{(pid=61970) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1161.892212] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Reconfiguring VM instance instance-00000068 to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1161.892464] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c0a210c-7d17-42ce-8a40-2905bfb0195d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.910101] env[61970]: DEBUG oslo_vmware.api [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1161.910101] env[61970]: value = "task-1356262" [ 1161.910101] env[61970]: _type = "Task" [ 1161.910101] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.917554] env[61970]: DEBUG oslo_vmware.api [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356262, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.419881] env[61970]: DEBUG oslo_vmware.api [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356262, 'name': ReconfigVM_Task, 'duration_secs': 0.206706} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.420227] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Reconfigured VM instance instance-00000068 to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1162.424752] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6abbab09-30a0-4075-8b46-ce7e75c27cbc {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.439413] env[61970]: DEBUG oslo_vmware.api [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1162.439413] env[61970]: value = "task-1356263" [ 1162.439413] env[61970]: _type = "Task" [ 1162.439413] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.446720] env[61970]: DEBUG oslo_vmware.api [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356263, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.948842] env[61970]: DEBUG oslo_vmware.api [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356263, 'name': ReconfigVM_Task, 'duration_secs': 0.132567} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.950078] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288890', 'volume_id': '9cb12dfc-bdee-43a7-b4ba-16190814effd', 'name': 'volume-9cb12dfc-bdee-43a7-b4ba-16190814effd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '5a8cdb36-1845-4806-a6bc-22afb47886e2', 'attached_at': '2024-10-24T13:48:27.000000', 'detached_at': '', 'volume_id': '9cb12dfc-bdee-43a7-b4ba-16190814effd', 'serial': '9cb12dfc-bdee-43a7-b4ba-16190814effd'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1162.950078] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1162.950521] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bfac265-93a3-4e67-bd05-bcba3ae56858 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.956617] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1162.956839] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-91be0e42-ac7c-4995-8885-19956d555c33 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.018794] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1163.019027] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1163.019223] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Deleting the datastore file [datastore2] 5a8cdb36-1845-4806-a6bc-22afb47886e2 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1163.019493] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ed98b40a-0b1b-4bbd-bf1d-214a85cb54ef {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.025199] env[61970]: DEBUG oslo_vmware.api [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1163.025199] env[61970]: value = "task-1356265" [ 1163.025199] env[61970]: _type = "Task" [ 1163.025199] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.032809] env[61970]: DEBUG oslo_vmware.api [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356265, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.535210] env[61970]: DEBUG oslo_vmware.api [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356265, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.422524} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.535573] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1163.535714] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1163.535810] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1163.536026] env[61970]: INFO nova.compute.manager [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Took 2.23 seconds to destroy the instance on the hypervisor. [ 1163.536289] env[61970]: DEBUG oslo.service.loopingcall [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1163.536484] env[61970]: DEBUG nova.compute.manager [-] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1163.536584] env[61970]: DEBUG nova.network.neutron [-] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1163.937448] env[61970]: DEBUG nova.compute.manager [req-af939f61-0a9f-4a63-af1f-12dc0a94f6ec req-00e29d2a-3742-43ad-9e37-136f305949ff service nova] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Received event network-vif-deleted-696b0179-1272-4ae7-ab15-727fe4682c42 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1163.937550] env[61970]: INFO nova.compute.manager [req-af939f61-0a9f-4a63-af1f-12dc0a94f6ec req-00e29d2a-3742-43ad-9e37-136f305949ff service nova] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Neutron deleted interface 696b0179-1272-4ae7-ab15-727fe4682c42; detaching it from the instance and deleting it from the info cache [ 1163.937730] env[61970]: DEBUG nova.network.neutron [req-af939f61-0a9f-4a63-af1f-12dc0a94f6ec req-00e29d2a-3742-43ad-9e37-136f305949ff service nova] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.417890] env[61970]: DEBUG nova.network.neutron [-] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.439865] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c7a9c02e-2fe0-4924-a977-d2aa7481fe1a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.450057] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d660954-bac0-4147-8a7f-5e24ee422b5a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.476310] env[61970]: DEBUG nova.compute.manager [req-af939f61-0a9f-4a63-af1f-12dc0a94f6ec req-00e29d2a-3742-43ad-9e37-136f305949ff service nova] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Detach interface failed, port_id=696b0179-1272-4ae7-ab15-727fe4682c42, reason: Instance 5a8cdb36-1845-4806-a6bc-22afb47886e2 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1164.923083] env[61970]: INFO nova.compute.manager [-] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Took 1.39 seconds to deallocate network for instance. [ 1165.094252] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.094506] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.094759] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.094964] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.095157] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.097330] env[61970]: INFO nova.compute.manager [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Terminating instance [ 1165.099150] env[61970]: DEBUG nova.compute.manager [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1165.099366] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1165.099647] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49afdd3d-fb19-4a8e-88e0-c0884d259474 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.107513] env[61970]: DEBUG oslo_vmware.api [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1165.107513] env[61970]: value = "task-1356266" [ 1165.107513] env[61970]: _type = "Task" [ 1165.107513] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.115501] env[61970]: DEBUG oslo_vmware.api [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356266, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.468653] env[61970]: INFO nova.compute.manager [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Took 0.55 seconds to detach 1 volumes for instance. [ 1165.618077] env[61970]: DEBUG oslo_vmware.api [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356266, 'name': PowerOffVM_Task, 'duration_secs': 0.288287} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.618367] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1165.618566] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Volume detach. Driver type: vmdk {{(pid=61970) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1165.618761] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288883', 'volume_id': 'ed3d65b8-03da-4484-8a98-b0e8211e6d54', 'name': 'volume-ed3d65b8-03da-4484-8a98-b0e8211e6d54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8', 'attached_at': '2024-10-24T13:47:58.000000', 'detached_at': '', 'volume_id': 'ed3d65b8-03da-4484-8a98-b0e8211e6d54', 'serial': 'ed3d65b8-03da-4484-8a98-b0e8211e6d54'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1165.619595] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9845b899-bb6c-4d0f-b5f5-6398ba4c6ff0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.637283] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b1b60d4-7c58-4c6e-a381-7952c78147f4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.643254] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-087a7e4b-accb-4c0c-bd60-e239a9d90282 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.659294] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908a9095-1d86-4210-af87-f35bc643feef {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.674230] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] The volume has not been displaced from its original location: [datastore1] volume-ed3d65b8-03da-4484-8a98-b0e8211e6d54/volume-ed3d65b8-03da-4484-8a98-b0e8211e6d54.vmdk. No consolidation needed. {{(pid=61970) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1165.679287] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Reconfiguring VM instance instance-0000006a to detach disk 2000 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1165.679558] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16129c6b-7b29-4fe0-b821-f47a7616e15a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.696372] env[61970]: DEBUG oslo_vmware.api [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1165.696372] env[61970]: value = "task-1356267" [ 1165.696372] env[61970]: _type = "Task" [ 1165.696372] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.703703] env[61970]: DEBUG oslo_vmware.api [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356267, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.974763] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.975152] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.975300] env[61970]: DEBUG nova.objects.instance [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lazy-loading 'resources' on Instance uuid 5a8cdb36-1845-4806-a6bc-22afb47886e2 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1166.206424] env[61970]: DEBUG oslo_vmware.api [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356267, 'name': ReconfigVM_Task, 'duration_secs': 0.161505} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.206708] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Reconfigured VM instance instance-0000006a to detach disk 2000 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1166.211347] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6afc1c1b-4d01-4b10-abf9-6743116336b8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.226316] env[61970]: DEBUG oslo_vmware.api [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1166.226316] env[61970]: value = "task-1356268" [ 1166.226316] env[61970]: _type = "Task" [ 1166.226316] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.235022] env[61970]: DEBUG oslo_vmware.api [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356268, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.547454] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89238347-c58f-4512-8371-5ec1e4b06c82 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.555268] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-990eea41-a7f8-41ea-bb7b-a7717d15e498 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.584126] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36da03ce-da91-40b9-aa9d-160455f0a7cd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.591134] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-916326b8-e630-4a93-bea6-8d4972968c14 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.604050] env[61970]: DEBUG nova.compute.provider_tree [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1166.736872] env[61970]: DEBUG oslo_vmware.api [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356268, 'name': ReconfigVM_Task, 'duration_secs': 0.143319} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.737176] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288883', 'volume_id': 'ed3d65b8-03da-4484-8a98-b0e8211e6d54', 'name': 'volume-ed3d65b8-03da-4484-8a98-b0e8211e6d54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8', 'attached_at': '2024-10-24T13:47:58.000000', 'detached_at': '', 'volume_id': 'ed3d65b8-03da-4484-8a98-b0e8211e6d54', 'serial': 'ed3d65b8-03da-4484-8a98-b0e8211e6d54'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1166.737431] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1166.738814] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25cd90ae-dfce-4ac8-ba27-96941af739ec {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.744231] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1166.744451] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f37b0bed-a69f-49ab-8213-5a1dac284a55 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.809459] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1166.809760] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1166.809950] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleting the datastore file [datastore1] a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1166.810236] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-adcfe25e-d343-4484-acd6-5917c0ef1d39 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.818163] env[61970]: DEBUG oslo_vmware.api [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1166.818163] env[61970]: value = "task-1356270" [ 1166.818163] env[61970]: _type = "Task" [ 1166.818163] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.825246] env[61970]: DEBUG oslo_vmware.api [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356270, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.107287] env[61970]: DEBUG nova.scheduler.client.report [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1167.331475] env[61970]: DEBUG oslo_vmware.api [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356270, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.093422} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.331855] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1167.332148] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1167.332410] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1167.332661] env[61970]: INFO nova.compute.manager [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Took 2.23 seconds to destroy the instance on the hypervisor. [ 1167.332994] env[61970]: DEBUG oslo.service.loopingcall [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1167.333265] env[61970]: DEBUG nova.compute.manager [-] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1167.333390] env[61970]: DEBUG nova.network.neutron [-] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1167.612245] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.637s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.630357] env[61970]: INFO nova.scheduler.client.report [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Deleted allocations for instance 5a8cdb36-1845-4806-a6bc-22afb47886e2 [ 1167.780566] env[61970]: DEBUG nova.compute.manager [req-be0fb47f-e2ea-4911-80fc-87ba4385e4b2 req-59ec6feb-7d74-4789-b743-68584c30b86e service nova] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Received event network-vif-deleted-8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1167.780868] env[61970]: INFO nova.compute.manager [req-be0fb47f-e2ea-4911-80fc-87ba4385e4b2 req-59ec6feb-7d74-4789-b743-68584c30b86e service nova] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Neutron deleted interface 8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d; detaching it from the instance and deleting it from the info cache [ 1167.781114] env[61970]: DEBUG nova.network.neutron [req-be0fb47f-e2ea-4911-80fc-87ba4385e4b2 req-59ec6feb-7d74-4789-b743-68584c30b86e service nova] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1168.137362] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4112e7a6-a6f9-462f-8f81-6fbe87992b2c tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "5a8cdb36-1845-4806-a6bc-22afb47886e2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.835s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.258885] env[61970]: DEBUG nova.network.neutron [-] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1168.284562] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2b04358d-0505-4ab8-af6c-df184ef81ca7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.295255] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f633bd91-50c2-4aae-9099-c113b203f5dd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.319695] env[61970]: DEBUG nova.compute.manager [req-be0fb47f-e2ea-4911-80fc-87ba4385e4b2 req-59ec6feb-7d74-4789-b743-68584c30b86e service nova] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Detach interface failed, port_id=8d2476c8-b7e5-47e8-a8ec-657a65f5ec1d, reason: Instance a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1168.761729] env[61970]: INFO nova.compute.manager [-] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Took 1.43 seconds to deallocate network for instance. [ 1169.301563] env[61970]: INFO nova.compute.manager [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Took 0.54 seconds to detach 1 volumes for instance. [ 1169.303727] env[61970]: DEBUG nova.compute.manager [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Deleting volume: ed3d65b8-03da-4484-8a98-b0e8211e6d54 {{(pid=61970) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1169.487926] env[61970]: DEBUG oslo_concurrency.lockutils [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "bead8709-273f-433f-9d06-4c01510aaca1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.488188] env[61970]: DEBUG oslo_concurrency.lockutils [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "bead8709-273f-433f-9d06-4c01510aaca1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.844612] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.844908] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.845158] env[61970]: DEBUG nova.objects.instance [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lazy-loading 'resources' on Instance uuid a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1169.991948] env[61970]: DEBUG nova.compute.manager [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1170.413536] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4d48ec-62bf-4394-b26f-a6ae3fffb71b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.420989] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3bb681c-7a45-4caf-aa51-51e2e179bc2e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.449575] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc948b9e-288b-4524-a7f7-05d3ec228c6d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.456520] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c5b84a-5d39-4e21-9d61-339ff59235ce {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.469094] env[61970]: DEBUG nova.compute.provider_tree [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1170.511664] env[61970]: DEBUG oslo_concurrency.lockutils [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.972195] env[61970]: DEBUG nova.scheduler.client.report [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1171.477011] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.632s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.479320] env[61970]: DEBUG oslo_concurrency.lockutils [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.968s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.481228] env[61970]: INFO nova.compute.claims [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1171.494615] env[61970]: INFO nova.scheduler.client.report [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleted allocations for instance a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8 [ 1172.001833] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ff115c48-4b88-48aa-b39b-d750203c9f9d tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.907s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.546578] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3334e83b-7c5a-4938-a473-ad27f1098b61 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.554301] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40e3ffce-3a3c-419e-ab1d-5b94a2eb4fea {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.582556] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84de5be0-63f1-4cf6-b978-fad25c1db04e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.589104] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132a78e0-bec0-492d-8fe4-7eea1a395cd9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.602625] env[61970]: DEBUG nova.compute.provider_tree [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1172.727628] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "183133c6-7e4e-4dd4-b8e0-cb4f8338264c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.727867] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "183133c6-7e4e-4dd4-b8e0-cb4f8338264c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.728098] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "183133c6-7e4e-4dd4-b8e0-cb4f8338264c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.728294] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "183133c6-7e4e-4dd4-b8e0-cb4f8338264c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.728465] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "183133c6-7e4e-4dd4-b8e0-cb4f8338264c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.730536] env[61970]: INFO nova.compute.manager [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Terminating instance [ 1172.732253] env[61970]: DEBUG nova.compute.manager [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1172.732452] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1172.733239] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b1e700e-e9df-4cea-b982-f98e5c7669b1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.741193] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1172.741193] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80d7a84d-d5c2-4f3d-9735-2ca945e5ae89 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.749026] env[61970]: DEBUG oslo_vmware.api [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1172.749026] env[61970]: value = "task-1356272" [ 1172.749026] env[61970]: _type = "Task" [ 1172.749026] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.755759] env[61970]: DEBUG oslo_vmware.api [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356272, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.106248] env[61970]: DEBUG nova.scheduler.client.report [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1173.257393] env[61970]: DEBUG oslo_vmware.api [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356272, 'name': PowerOffVM_Task, 'duration_secs': 0.179728} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.257672] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1173.258228] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1173.258228] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f54c998d-bca3-46c7-aa92-93f10b3d2ff6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.320293] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1173.320519] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1173.320771] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleting the datastore file [datastore2] 183133c6-7e4e-4dd4-b8e0-cb4f8338264c {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1173.321352] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b1154cb7-407e-49d4-a837-cc3e56f15648 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.327614] env[61970]: DEBUG oslo_vmware.api [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1173.327614] env[61970]: value = "task-1356274" [ 1173.327614] env[61970]: _type = "Task" [ 1173.327614] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.335275] env[61970]: DEBUG oslo_vmware.api [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356274, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.611427] env[61970]: DEBUG oslo_concurrency.lockutils [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.132s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1173.612028] env[61970]: DEBUG nova.compute.manager [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1173.838136] env[61970]: DEBUG oslo_vmware.api [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356274, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136351} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.838331] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1173.838510] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1173.838692] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1173.838871] env[61970]: INFO nova.compute.manager [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1173.839136] env[61970]: DEBUG oslo.service.loopingcall [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1173.839349] env[61970]: DEBUG nova.compute.manager [-] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1173.839456] env[61970]: DEBUG nova.network.neutron [-] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1174.085008] env[61970]: DEBUG nova.compute.manager [req-809ea727-1048-4b61-894c-467d9456ae16 req-31f7d83d-522e-4044-8e9f-78d55cc4e1c2 service nova] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Received event network-vif-deleted-f89f6ca5-913c-4eff-ab7e-2199ee90ccd7 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1174.085213] env[61970]: INFO nova.compute.manager [req-809ea727-1048-4b61-894c-467d9456ae16 req-31f7d83d-522e-4044-8e9f-78d55cc4e1c2 service nova] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Neutron deleted interface f89f6ca5-913c-4eff-ab7e-2199ee90ccd7; detaching it from the instance and deleting it from the info cache [ 1174.085401] env[61970]: DEBUG nova.network.neutron [req-809ea727-1048-4b61-894c-467d9456ae16 req-31f7d83d-522e-4044-8e9f-78d55cc4e1c2 service nova] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1174.117225] env[61970]: DEBUG nova.compute.utils [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1174.121880] env[61970]: DEBUG nova.compute.manager [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1174.121880] env[61970]: DEBUG nova.network.neutron [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1174.160420] env[61970]: DEBUG nova.policy [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2b8322c5de6483aaf032781d28e950d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09033ee688384ca287b4fdad6e67cb1f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 1174.450733] env[61970]: DEBUG nova.network.neutron [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Successfully created port: be465191-5965-4079-a636-95bfbd16dd1d {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1174.562444] env[61970]: DEBUG nova.network.neutron [-] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1174.589037] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5c1984c8-d1b8-4750-bfa7-f8696eb29333 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.600507] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea8ec9c-91db-4993-9e6c-b6f57837eaa1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.626224] env[61970]: DEBUG nova.compute.manager [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1174.628856] env[61970]: DEBUG nova.compute.manager [req-809ea727-1048-4b61-894c-467d9456ae16 req-31f7d83d-522e-4044-8e9f-78d55cc4e1c2 service nova] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Detach interface failed, port_id=f89f6ca5-913c-4eff-ab7e-2199ee90ccd7, reason: Instance 183133c6-7e4e-4dd4-b8e0-cb4f8338264c could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1175.064789] env[61970]: INFO nova.compute.manager [-] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Took 1.23 seconds to deallocate network for instance. [ 1175.573741] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1175.574043] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.574272] env[61970]: DEBUG nova.objects.instance [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lazy-loading 'resources' on Instance uuid 183133c6-7e4e-4dd4-b8e0-cb4f8338264c {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1175.635928] env[61970]: DEBUG nova.compute.manager [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1175.660977] env[61970]: DEBUG nova.virt.hardware [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1175.661255] env[61970]: DEBUG nova.virt.hardware [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1175.661424] env[61970]: DEBUG nova.virt.hardware [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1175.661597] env[61970]: DEBUG nova.virt.hardware [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1175.661749] env[61970]: DEBUG nova.virt.hardware [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1175.661960] env[61970]: DEBUG nova.virt.hardware [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1175.662148] env[61970]: DEBUG nova.virt.hardware [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1175.662303] env[61970]: DEBUG nova.virt.hardware [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1175.662477] env[61970]: DEBUG nova.virt.hardware [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1175.662645] env[61970]: DEBUG nova.virt.hardware [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1175.662819] env[61970]: DEBUG nova.virt.hardware [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1175.663697] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d1993b6-7db6-4b57-9f9f-dc5f5f904bc5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.671769] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e615dbf3-ae4f-4028-941d-5886e6017d6d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.909039] env[61970]: DEBUG nova.network.neutron [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Successfully updated port: be465191-5965-4079-a636-95bfbd16dd1d {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1176.109970] env[61970]: DEBUG nova.compute.manager [req-8a4e37de-065d-423e-b4e1-c00156121cb0 req-69e5448b-c31c-474f-b347-d611bc5d74d8 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Received event network-vif-plugged-be465191-5965-4079-a636-95bfbd16dd1d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1176.110208] env[61970]: DEBUG oslo_concurrency.lockutils [req-8a4e37de-065d-423e-b4e1-c00156121cb0 req-69e5448b-c31c-474f-b347-d611bc5d74d8 service nova] Acquiring lock "bead8709-273f-433f-9d06-4c01510aaca1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.110416] env[61970]: DEBUG oslo_concurrency.lockutils [req-8a4e37de-065d-423e-b4e1-c00156121cb0 req-69e5448b-c31c-474f-b347-d611bc5d74d8 service nova] Lock "bead8709-273f-433f-9d06-4c01510aaca1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.110620] env[61970]: DEBUG oslo_concurrency.lockutils [req-8a4e37de-065d-423e-b4e1-c00156121cb0 req-69e5448b-c31c-474f-b347-d611bc5d74d8 service nova] Lock "bead8709-273f-433f-9d06-4c01510aaca1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.110787] env[61970]: DEBUG nova.compute.manager [req-8a4e37de-065d-423e-b4e1-c00156121cb0 req-69e5448b-c31c-474f-b347-d611bc5d74d8 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] No waiting events found dispatching network-vif-plugged-be465191-5965-4079-a636-95bfbd16dd1d {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1176.110953] env[61970]: WARNING nova.compute.manager [req-8a4e37de-065d-423e-b4e1-c00156121cb0 req-69e5448b-c31c-474f-b347-d611bc5d74d8 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Received unexpected event network-vif-plugged-be465191-5965-4079-a636-95bfbd16dd1d for instance with vm_state building and task_state spawning. [ 1176.111140] env[61970]: DEBUG nova.compute.manager [req-8a4e37de-065d-423e-b4e1-c00156121cb0 req-69e5448b-c31c-474f-b347-d611bc5d74d8 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Received event network-changed-be465191-5965-4079-a636-95bfbd16dd1d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1176.111297] env[61970]: DEBUG nova.compute.manager [req-8a4e37de-065d-423e-b4e1-c00156121cb0 req-69e5448b-c31c-474f-b347-d611bc5d74d8 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Refreshing instance network info cache due to event network-changed-be465191-5965-4079-a636-95bfbd16dd1d. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1176.111479] env[61970]: DEBUG oslo_concurrency.lockutils [req-8a4e37de-065d-423e-b4e1-c00156121cb0 req-69e5448b-c31c-474f-b347-d611bc5d74d8 service nova] Acquiring lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1176.111616] env[61970]: DEBUG oslo_concurrency.lockutils [req-8a4e37de-065d-423e-b4e1-c00156121cb0 req-69e5448b-c31c-474f-b347-d611bc5d74d8 service nova] Acquired lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.111780] env[61970]: DEBUG nova.network.neutron [req-8a4e37de-065d-423e-b4e1-c00156121cb0 req-69e5448b-c31c-474f-b347-d611bc5d74d8 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Refreshing network info cache for port be465191-5965-4079-a636-95bfbd16dd1d {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1176.137320] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8e7368-b30a-4e6d-a9a2-aaf0b1d8d69d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.144893] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a42b4a4-b266-4cb5-bd12-a8bc20a633d3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.175489] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbeba8ed-62da-4730-adae-b0488748d6fd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.182673] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c35b706-ddfd-478c-af20-cf50bd0e39cd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.195342] env[61970]: DEBUG nova.compute.provider_tree [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1176.411534] env[61970]: DEBUG oslo_concurrency.lockutils [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1176.640948] env[61970]: DEBUG nova.network.neutron [req-8a4e37de-065d-423e-b4e1-c00156121cb0 req-69e5448b-c31c-474f-b347-d611bc5d74d8 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1176.698785] env[61970]: DEBUG nova.scheduler.client.report [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1176.711693] env[61970]: DEBUG nova.network.neutron [req-8a4e37de-065d-423e-b4e1-c00156121cb0 req-69e5448b-c31c-474f-b347-d611bc5d74d8 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1177.203471] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.629s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.213899] env[61970]: DEBUG oslo_concurrency.lockutils [req-8a4e37de-065d-423e-b4e1-c00156121cb0 req-69e5448b-c31c-474f-b347-d611bc5d74d8 service nova] Releasing lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1177.214834] env[61970]: DEBUG oslo_concurrency.lockutils [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1177.214834] env[61970]: DEBUG nova.network.neutron [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1177.224855] env[61970]: INFO nova.scheduler.client.report [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleted allocations for instance 183133c6-7e4e-4dd4-b8e0-cb4f8338264c [ 1177.732474] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f9411920-ec5c-49bb-ac62-822b21af46cf tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "183133c6-7e4e-4dd4-b8e0-cb4f8338264c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.004s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.746089] env[61970]: DEBUG nova.network.neutron [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1177.876429] env[61970]: DEBUG nova.network.neutron [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Updating instance_info_cache with network_info: [{"id": "be465191-5965-4079-a636-95bfbd16dd1d", "address": "fa:16:3e:b0:0a:ef", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe465191-59", "ovs_interfaceid": "be465191-5965-4079-a636-95bfbd16dd1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1178.378792] env[61970]: DEBUG oslo_concurrency.lockutils [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1178.379079] env[61970]: DEBUG nova.compute.manager [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Instance network_info: |[{"id": "be465191-5965-4079-a636-95bfbd16dd1d", "address": "fa:16:3e:b0:0a:ef", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe465191-59", "ovs_interfaceid": "be465191-5965-4079-a636-95bfbd16dd1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1178.379563] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:0a:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '298bb8ef-4765-494c-b157-7a349218bd1e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'be465191-5965-4079-a636-95bfbd16dd1d', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1178.386947] env[61970]: DEBUG oslo.service.loopingcall [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1178.387176] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1178.387402] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-097e3142-20c2-4958-acef-8318ee8caf99 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.411587] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1178.411587] env[61970]: value = "task-1356276" [ 1178.411587] env[61970]: _type = "Task" [ 1178.411587] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.418803] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356276, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.921379] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356276, 'name': CreateVM_Task, 'duration_secs': 0.317948} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.921744] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1178.922215] env[61970]: DEBUG oslo_concurrency.lockutils [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.922385] env[61970]: DEBUG oslo_concurrency.lockutils [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.922713] env[61970]: DEBUG oslo_concurrency.lockutils [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1178.922956] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b70d4b6-c61a-4423-90d0-727e0bb097d2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.927466] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1178.927466] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d3f676-3705-f896-702e-950cbc0dba1c" [ 1178.927466] env[61970]: _type = "Task" [ 1178.927466] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.935150] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d3f676-3705-f896-702e-950cbc0dba1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.437109] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d3f676-3705-f896-702e-950cbc0dba1c, 'name': SearchDatastore_Task, 'duration_secs': 0.012445} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.437422] env[61970]: DEBUG oslo_concurrency.lockutils [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.437695] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1179.437938] env[61970]: DEBUG oslo_concurrency.lockutils [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.438111] env[61970]: DEBUG oslo_concurrency.lockutils [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.438289] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1179.438543] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97cf82a4-1a66-4e4d-a387-8b91b17de4eb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.446401] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1179.446575] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1179.447255] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9f4bb48-1a36-49ee-ad17-b7a2585ed3f4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.451780] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1179.451780] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5209ce80-3372-343a-2e49-316dbcd8225b" [ 1179.451780] env[61970]: _type = "Task" [ 1179.451780] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.458873] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5209ce80-3372-343a-2e49-316dbcd8225b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.962104] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5209ce80-3372-343a-2e49-316dbcd8225b, 'name': SearchDatastore_Task, 'duration_secs': 0.008325} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.962851] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71912154-799c-4256-88c1-42ea69d999d5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.967684] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1179.967684] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523d1791-6a21-f061-9905-69f3238d3577" [ 1179.967684] env[61970]: _type = "Task" [ 1179.967684] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.974974] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523d1791-6a21-f061-9905-69f3238d3577, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.477732] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523d1791-6a21-f061-9905-69f3238d3577, 'name': SearchDatastore_Task, 'duration_secs': 0.009129} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.477996] env[61970]: DEBUG oslo_concurrency.lockutils [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1180.478268] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] bead8709-273f-433f-9d06-4c01510aaca1/bead8709-273f-433f-9d06-4c01510aaca1.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1180.478522] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-40c932e0-2cd8-4c69-91eb-7387ecb78e53 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.484807] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1180.484807] env[61970]: value = "task-1356277" [ 1180.484807] env[61970]: _type = "Task" [ 1180.484807] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.492348] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356277, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.994204] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356277, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442801} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.994533] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] bead8709-273f-433f-9d06-4c01510aaca1/bead8709-273f-433f-9d06-4c01510aaca1.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1180.994606] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1180.994850] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-822fffd4-0119-44ef-a538-c3dac82ba7f4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.000471] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1181.000471] env[61970]: value = "task-1356278" [ 1181.000471] env[61970]: _type = "Task" [ 1181.000471] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.008068] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356278, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.329029] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "0304640a-6c46-4f57-a8b8-5058da2e3de4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1181.329253] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "0304640a-6c46-4f57-a8b8-5058da2e3de4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1181.329516] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "0304640a-6c46-4f57-a8b8-5058da2e3de4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1181.329728] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "0304640a-6c46-4f57-a8b8-5058da2e3de4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1181.329907] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "0304640a-6c46-4f57-a8b8-5058da2e3de4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1181.332136] env[61970]: INFO nova.compute.manager [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Terminating instance [ 1181.333835] env[61970]: DEBUG nova.compute.manager [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1181.334048] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1181.334848] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb5367ce-e309-476a-b57e-bd303231d9f3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.343446] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1181.343667] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8a44a16-6722-4282-8071-df4ff9eb38ff {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.349256] env[61970]: DEBUG oslo_vmware.api [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1181.349256] env[61970]: value = "task-1356279" [ 1181.349256] env[61970]: _type = "Task" [ 1181.349256] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.357367] env[61970]: DEBUG oslo_vmware.api [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356279, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.510939] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356278, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058737} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.511239] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1181.512248] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-565fc430-9f1f-4375-a972-a9cb5f28e939 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.534670] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] bead8709-273f-433f-9d06-4c01510aaca1/bead8709-273f-433f-9d06-4c01510aaca1.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1181.534955] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e02c939a-b35e-4acb-9648-f6042fc28e23 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.553483] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1181.553483] env[61970]: value = "task-1356280" [ 1181.553483] env[61970]: _type = "Task" [ 1181.553483] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.561146] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356280, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.859311] env[61970]: DEBUG oslo_vmware.api [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356279, 'name': PowerOffVM_Task, 'duration_secs': 0.18778} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.859607] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1181.859778] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1181.860386] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-615dd9e8-f645-4135-87d2-ecf3d692b4f3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.927048] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1181.927048] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1181.927048] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleting the datastore file [datastore1] 0304640a-6c46-4f57-a8b8-5058da2e3de4 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1181.927315] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-811b45f6-74af-44ed-a53b-f8059f67e632 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.933866] env[61970]: DEBUG oslo_vmware.api [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for the task: (returnval){ [ 1181.933866] env[61970]: value = "task-1356282" [ 1181.933866] env[61970]: _type = "Task" [ 1181.933866] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.941362] env[61970]: DEBUG oslo_vmware.api [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356282, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.064313] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356280, 'name': ReconfigVM_Task, 'duration_secs': 0.258952} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.064773] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Reconfigured VM instance instance-0000006b to attach disk [datastore2] bead8709-273f-433f-9d06-4c01510aaca1/bead8709-273f-433f-9d06-4c01510aaca1.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1182.065286] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-993517b9-f6d5-4ce8-8749-d1f7d97f0d97 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.071391] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1182.071391] env[61970]: value = "task-1356283" [ 1182.071391] env[61970]: _type = "Task" [ 1182.071391] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.079555] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356283, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.444033] env[61970]: DEBUG oslo_vmware.api [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Task: {'id': task-1356282, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1487} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.444299] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1182.444483] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1182.444666] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1182.444842] env[61970]: INFO nova.compute.manager [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1182.445098] env[61970]: DEBUG oslo.service.loopingcall [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1182.445293] env[61970]: DEBUG nova.compute.manager [-] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1182.445389] env[61970]: DEBUG nova.network.neutron [-] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1182.580915] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356283, 'name': Rename_Task, 'duration_secs': 0.138424} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.581201] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1182.581470] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-05adc4e0-3eaa-4c12-8eaa-14f83e91e75f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.587758] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1182.587758] env[61970]: value = "task-1356284" [ 1182.587758] env[61970]: _type = "Task" [ 1182.587758] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.594972] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356284, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.692691] env[61970]: DEBUG nova.compute.manager [req-d7aa2cb5-2480-4a1f-824b-3549477af9ab req-3ebdfb72-567c-47ff-8fb5-5a77bfcfa7a2 service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Received event network-vif-deleted-2545e2b0-a432-44ae-a287-0b68ceab7031 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1182.693055] env[61970]: INFO nova.compute.manager [req-d7aa2cb5-2480-4a1f-824b-3549477af9ab req-3ebdfb72-567c-47ff-8fb5-5a77bfcfa7a2 service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Neutron deleted interface 2545e2b0-a432-44ae-a287-0b68ceab7031; detaching it from the instance and deleting it from the info cache [ 1182.693359] env[61970]: DEBUG nova.network.neutron [req-d7aa2cb5-2480-4a1f-824b-3549477af9ab req-3ebdfb72-567c-47ff-8fb5-5a77bfcfa7a2 service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1183.099256] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356284, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.169867] env[61970]: DEBUG nova.network.neutron [-] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1183.196056] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ae7f1b3e-923f-429a-ae5b-0b772a16bb8e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.205233] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40952bb9-c979-48ea-9b94-443abe03bd36 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.228784] env[61970]: DEBUG nova.compute.manager [req-d7aa2cb5-2480-4a1f-824b-3549477af9ab req-3ebdfb72-567c-47ff-8fb5-5a77bfcfa7a2 service nova] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Detach interface failed, port_id=2545e2b0-a432-44ae-a287-0b68ceab7031, reason: Instance 0304640a-6c46-4f57-a8b8-5058da2e3de4 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1183.598413] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356284, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.672205] env[61970]: INFO nova.compute.manager [-] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Took 1.23 seconds to deallocate network for instance. [ 1184.098124] env[61970]: DEBUG oslo_vmware.api [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356284, 'name': PowerOnVM_Task, 'duration_secs': 1.110169} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.098379] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1184.098587] env[61970]: INFO nova.compute.manager [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Took 8.46 seconds to spawn the instance on the hypervisor. [ 1184.098783] env[61970]: DEBUG nova.compute.manager [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1184.099557] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70880634-f9b3-4a07-9282-a99ad801a7c1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.179740] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.179827] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.180040] env[61970]: DEBUG nova.objects.instance [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lazy-loading 'resources' on Instance uuid 0304640a-6c46-4f57-a8b8-5058da2e3de4 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1184.616024] env[61970]: INFO nova.compute.manager [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Took 14.12 seconds to build instance. [ 1184.733998] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d05255a2-e7f6-487a-86e4-695001aef784 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.741806] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fcac08a-fbc8-48c2-8392-9ac9993a9864 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.772225] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e96e8a24-03bd-4aac-9e7f-ffa4a71ceeb7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.779323] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f65637f9-519f-47c5-8c66-1fa7cad56f20 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.792397] env[61970]: DEBUG nova.compute.provider_tree [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1184.913314] env[61970]: DEBUG nova.compute.manager [req-6c46ddc9-3a0c-464f-b7b2-91a44fae61c4 req-2c80fa89-e936-41ae-be7f-abfddd4a07a7 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Received event network-changed-be465191-5965-4079-a636-95bfbd16dd1d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1184.913520] env[61970]: DEBUG nova.compute.manager [req-6c46ddc9-3a0c-464f-b7b2-91a44fae61c4 req-2c80fa89-e936-41ae-be7f-abfddd4a07a7 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Refreshing instance network info cache due to event network-changed-be465191-5965-4079-a636-95bfbd16dd1d. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1184.913785] env[61970]: DEBUG oslo_concurrency.lockutils [req-6c46ddc9-3a0c-464f-b7b2-91a44fae61c4 req-2c80fa89-e936-41ae-be7f-abfddd4a07a7 service nova] Acquiring lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1184.913875] env[61970]: DEBUG oslo_concurrency.lockutils [req-6c46ddc9-3a0c-464f-b7b2-91a44fae61c4 req-2c80fa89-e936-41ae-be7f-abfddd4a07a7 service nova] Acquired lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1184.914040] env[61970]: DEBUG nova.network.neutron [req-6c46ddc9-3a0c-464f-b7b2-91a44fae61c4 req-2c80fa89-e936-41ae-be7f-abfddd4a07a7 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Refreshing network info cache for port be465191-5965-4079-a636-95bfbd16dd1d {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1185.117236] env[61970]: DEBUG oslo_concurrency.lockutils [None req-baddb47d-7bc7-42f8-95c6-45dcce03d50b tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "bead8709-273f-433f-9d06-4c01510aaca1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.629s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.296111] env[61970]: DEBUG nova.scheduler.client.report [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1185.619183] env[61970]: DEBUG nova.network.neutron [req-6c46ddc9-3a0c-464f-b7b2-91a44fae61c4 req-2c80fa89-e936-41ae-be7f-abfddd4a07a7 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Updated VIF entry in instance network info cache for port be465191-5965-4079-a636-95bfbd16dd1d. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1185.619632] env[61970]: DEBUG nova.network.neutron [req-6c46ddc9-3a0c-464f-b7b2-91a44fae61c4 req-2c80fa89-e936-41ae-be7f-abfddd4a07a7 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Updating instance_info_cache with network_info: [{"id": "be465191-5965-4079-a636-95bfbd16dd1d", "address": "fa:16:3e:b0:0a:ef", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe465191-59", "ovs_interfaceid": "be465191-5965-4079-a636-95bfbd16dd1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1185.800328] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.620s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.821652] env[61970]: INFO nova.scheduler.client.report [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Deleted allocations for instance 0304640a-6c46-4f57-a8b8-5058da2e3de4 [ 1186.122399] env[61970]: DEBUG oslo_concurrency.lockutils [req-6c46ddc9-3a0c-464f-b7b2-91a44fae61c4 req-2c80fa89-e936-41ae-be7f-abfddd4a07a7 service nova] Releasing lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1186.329753] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6272e5d5-d5b2-4812-94ae-22f770624176 tempest-ServerActionsTestOtherA-534102670 tempest-ServerActionsTestOtherA-534102670-project-member] Lock "0304640a-6c46-4f57-a8b8-5058da2e3de4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.638107] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.638507] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.638605] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.638796] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.638968] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.641169] env[61970]: INFO nova.compute.manager [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Terminating instance [ 1188.642875] env[61970]: DEBUG nova.compute.manager [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1188.643088] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1188.643950] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e72329f-6978-4505-bf8a-f752f27f9b03 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.651358] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1188.651581] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f81d236-f383-4b42-bb87-2d5ab2d80c27 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.658919] env[61970]: DEBUG oslo_vmware.api [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1188.658919] env[61970]: value = "task-1356285" [ 1188.658919] env[61970]: _type = "Task" [ 1188.658919] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.665901] env[61970]: DEBUG oslo_vmware.api [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356285, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.168448] env[61970]: DEBUG oslo_vmware.api [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356285, 'name': PowerOffVM_Task, 'duration_secs': 0.212853} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.168705] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1189.168879] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1189.169143] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7326f79b-9269-4c88-8ee5-8b46a5614c8e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.576851] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1189.577036] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1189.577208] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Deleting the datastore file [datastore1] 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1189.577504] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-90e927e8-b862-46ab-88f3-5837784decd6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.583320] env[61970]: DEBUG oslo_vmware.api [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for the task: (returnval){ [ 1189.583320] env[61970]: value = "task-1356287" [ 1189.583320] env[61970]: _type = "Task" [ 1189.583320] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.590868] env[61970]: DEBUG oslo_vmware.api [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356287, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.093030] env[61970]: DEBUG oslo_vmware.api [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Task: {'id': task-1356287, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137415} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.093425] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1190.093472] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1190.093645] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1190.093824] env[61970]: INFO nova.compute.manager [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Took 1.45 seconds to destroy the instance on the hypervisor. [ 1190.094079] env[61970]: DEBUG oslo.service.loopingcall [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1190.094276] env[61970]: DEBUG nova.compute.manager [-] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1190.094371] env[61970]: DEBUG nova.network.neutron [-] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1190.502308] env[61970]: DEBUG nova.compute.manager [req-7b7e1642-5a59-4eac-b297-ea72dd196dde req-4eb5322f-2612-4b42-a63e-c0bca7b5c05d service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Received event network-vif-deleted-2fa02dfe-2cb0-4a72-bc30-4f991339dcb9 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1190.502722] env[61970]: INFO nova.compute.manager [req-7b7e1642-5a59-4eac-b297-ea72dd196dde req-4eb5322f-2612-4b42-a63e-c0bca7b5c05d service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Neutron deleted interface 2fa02dfe-2cb0-4a72-bc30-4f991339dcb9; detaching it from the instance and deleting it from the info cache [ 1190.502722] env[61970]: DEBUG nova.network.neutron [req-7b7e1642-5a59-4eac-b297-ea72dd196dde req-4eb5322f-2612-4b42-a63e-c0bca7b5c05d service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1190.981121] env[61970]: DEBUG nova.network.neutron [-] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1191.004824] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f1b70c3b-8948-4006-a5ef-09a52b3a3733 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.014383] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f5828bc-6f31-47bb-826a-cb0df6345d1f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.036660] env[61970]: DEBUG nova.compute.manager [req-7b7e1642-5a59-4eac-b297-ea72dd196dde req-4eb5322f-2612-4b42-a63e-c0bca7b5c05d service nova] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Detach interface failed, port_id=2fa02dfe-2cb0-4a72-bc30-4f991339dcb9, reason: Instance 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1191.484701] env[61970]: INFO nova.compute.manager [-] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Took 1.39 seconds to deallocate network for instance. [ 1191.993669] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.993942] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.994193] env[61970]: DEBUG nova.objects.instance [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lazy-loading 'resources' on Instance uuid 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1192.537716] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b070eb6-e21c-4acd-9e37-a14240a8e122 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.545134] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b72ba69-8d1f-4a4d-af6c-a5199e9a91d8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.574894] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f36d0913-aa4e-4b48-b303-1e917a317d00 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.581699] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e12d253-452a-471a-b135-fd6225c3ad2f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.595774] env[61970]: DEBUG nova.compute.provider_tree [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1193.098166] env[61970]: DEBUG nova.scheduler.client.report [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1193.603233] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.609s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.622212] env[61970]: INFO nova.scheduler.client.report [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Deleted allocations for instance 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6 [ 1194.130479] env[61970]: DEBUG oslo_concurrency.lockutils [None req-60ec4387-e1f7-466d-bd1a-c2ba55e5a708 tempest-AttachVolumeShelveTestJSON-674312887 tempest-AttachVolumeShelveTestJSON-674312887-project-member] Lock "3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.492s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1204.260129] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1206.260237] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1209.260222] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1209.260645] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1209.260645] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61970) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1209.260807] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1209.764255] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.764486] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.764784] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.764889] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61970) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1209.765809] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2d0248-a9dc-4cd1-b68e-a8cb6059b90d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.774220] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0d9fb9-6dfd-4a7d-821a-d28364cf6f25 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.788802] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c9c866-faf2-4570-a55c-b6ab2db19b9c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.794957] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd7f891e-a862-490f-9d8b-7deb31fce7d2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.822249] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180703MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61970) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1209.822400] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.822588] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.847394] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance bead8709-273f-433f-9d06-4c01510aaca1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1210.847662] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1210.847742] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1210.872284] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeed76ef-9620-4b4e-833c-f560051d4769 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.879913] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e83cc20d-8728-4a6f-aac5-daecfaf2881d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.910086] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e284b02a-9832-40ea-b49d-73e971bcbd02 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.916917] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6e1228-b3e9-4004-849c-a1acc78462c2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.929387] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1211.431920] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1211.937592] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1211.937949] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.115s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.932710] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1212.932936] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1212.933099] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Starting heal instance info cache {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1214.943592] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Didn't find any instances for network info cache update. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1214.943873] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1214.944070] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1216.266944] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1222.696484] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "bead8709-273f-433f-9d06-4c01510aaca1" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.696881] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "bead8709-273f-433f-9d06-4c01510aaca1" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.696966] env[61970]: INFO nova.compute.manager [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Shelving [ 1223.205273] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1223.205597] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bac020cd-9273-48df-aeb5-318311eefe67 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.215587] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1223.215587] env[61970]: value = "task-1356289" [ 1223.215587] env[61970]: _type = "Task" [ 1223.215587] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.223447] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356289, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.725349] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356289, 'name': PowerOffVM_Task, 'duration_secs': 0.200107} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.725718] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1223.726331] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e0c5009-46c4-4713-8585-54b947ec4211 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.743798] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e516d4e2-e8c1-4281-b014-df4ba0d6fc07 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.254132] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Creating Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1224.254448] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2056a428-6cc5-4ff5-8c7b-03b838bd47fd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.262473] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1224.262473] env[61970]: value = "task-1356290" [ 1224.262473] env[61970]: _type = "Task" [ 1224.262473] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.270043] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356290, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.773314] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356290, 'name': CreateSnapshot_Task, 'duration_secs': 0.428468} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.773671] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Created Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1224.774313] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d427a5c-5219-4c38-b83b-b9c62c8a81b8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.291932] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Creating linked-clone VM from snapshot {{(pid=61970) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1225.292253] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7e1843e8-419a-42f6-894d-47929dcaef00 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.300955] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1225.300955] env[61970]: value = "task-1356291" [ 1225.300955] env[61970]: _type = "Task" [ 1225.300955] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.308562] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356291, 'name': CloneVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.812682] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356291, 'name': CloneVM_Task} progress is 94%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.311377] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356291, 'name': CloneVM_Task, 'duration_secs': 0.922337} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.311644] env[61970]: INFO nova.virt.vmwareapi.vmops [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Created linked-clone VM from snapshot [ 1226.312394] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78cb688f-c74f-4f82-a38d-52de33250977 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.319250] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Uploading image 052216c7-6b4d-4f15-bb1f-25aa97395548 {{(pid=61970) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1226.339308] env[61970]: DEBUG oslo_vmware.rw_handles [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1226.339308] env[61970]: value = "vm-288896" [ 1226.339308] env[61970]: _type = "VirtualMachine" [ 1226.339308] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1226.339581] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-46a07891-652b-423e-89f5-2465b5cb4a7b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.345814] env[61970]: DEBUG oslo_vmware.rw_handles [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lease: (returnval){ [ 1226.345814] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a907a2-8bc9-8ecc-adc4-d2177b2640eb" [ 1226.345814] env[61970]: _type = "HttpNfcLease" [ 1226.345814] env[61970]: } obtained for exporting VM: (result){ [ 1226.345814] env[61970]: value = "vm-288896" [ 1226.345814] env[61970]: _type = "VirtualMachine" [ 1226.345814] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1226.346097] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the lease: (returnval){ [ 1226.346097] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a907a2-8bc9-8ecc-adc4-d2177b2640eb" [ 1226.346097] env[61970]: _type = "HttpNfcLease" [ 1226.346097] env[61970]: } to be ready. {{(pid=61970) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1226.352474] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1226.352474] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a907a2-8bc9-8ecc-adc4-d2177b2640eb" [ 1226.352474] env[61970]: _type = "HttpNfcLease" [ 1226.352474] env[61970]: } is initializing. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1226.854526] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1226.854526] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a907a2-8bc9-8ecc-adc4-d2177b2640eb" [ 1226.854526] env[61970]: _type = "HttpNfcLease" [ 1226.854526] env[61970]: } is ready. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1226.855046] env[61970]: DEBUG oslo_vmware.rw_handles [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1226.855046] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a907a2-8bc9-8ecc-adc4-d2177b2640eb" [ 1226.855046] env[61970]: _type = "HttpNfcLease" [ 1226.855046] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1226.855526] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-323d5e40-0981-46ba-ab1b-2a9fc9ea1962 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.863124] env[61970]: DEBUG oslo_vmware.rw_handles [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525aca2e-48cf-0c75-6bec-123b2d45b601/disk-0.vmdk from lease info. {{(pid=61970) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1226.863383] env[61970]: DEBUG oslo_vmware.rw_handles [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525aca2e-48cf-0c75-6bec-123b2d45b601/disk-0.vmdk for reading. {{(pid=61970) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1226.948092] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-eb8f1dbb-1a41-44f9-8452-fc1f1bf1bca5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.675194] env[61970]: DEBUG oslo_vmware.rw_handles [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525aca2e-48cf-0c75-6bec-123b2d45b601/disk-0.vmdk. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1234.676220] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d922ed8c-6684-445b-8404-901172625adb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.682572] env[61970]: DEBUG oslo_vmware.rw_handles [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525aca2e-48cf-0c75-6bec-123b2d45b601/disk-0.vmdk is in state: ready. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1234.682755] env[61970]: ERROR oslo_vmware.rw_handles [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525aca2e-48cf-0c75-6bec-123b2d45b601/disk-0.vmdk due to incomplete transfer. [ 1234.682988] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f5eb3336-b1a8-4638-9916-4c687a20c447 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.692870] env[61970]: DEBUG oslo_vmware.rw_handles [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525aca2e-48cf-0c75-6bec-123b2d45b601/disk-0.vmdk. {{(pid=61970) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1234.693101] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Uploaded image 052216c7-6b4d-4f15-bb1f-25aa97395548 to the Glance image server {{(pid=61970) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1234.695676] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Destroying the VM {{(pid=61970) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1234.696087] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-20e07a52-e486-4e40-b16a-57fe7a04a021 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.701494] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1234.701494] env[61970]: value = "task-1356293" [ 1234.701494] env[61970]: _type = "Task" [ 1234.701494] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.709708] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356293, 'name': Destroy_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.211678] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356293, 'name': Destroy_Task, 'duration_secs': 0.482518} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.211925] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Destroyed the VM [ 1235.212209] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Deleting Snapshot of the VM instance {{(pid=61970) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1235.212464] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ba40ffb4-7bdf-42ee-9776-b1cf73fbb6a6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.218491] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1235.218491] env[61970]: value = "task-1356294" [ 1235.218491] env[61970]: _type = "Task" [ 1235.218491] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.225610] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356294, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.727719] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356294, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.228122] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356294, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.728747] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356294, 'name': RemoveSnapshot_Task, 'duration_secs': 1.245504} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.729080] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Deleted Snapshot of the VM instance {{(pid=61970) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1236.729296] env[61970]: DEBUG nova.compute.manager [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1236.730086] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd3e3111-561c-455f-a3e2-bf1c06766cb3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.241652] env[61970]: INFO nova.compute.manager [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Shelve offloading [ 1237.243278] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1237.243525] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e22bf772-56ee-4093-887e-67468071e00c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.251065] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1237.251065] env[61970]: value = "task-1356295" [ 1237.251065] env[61970]: _type = "Task" [ 1237.251065] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.259886] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] VM already powered off {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1237.260075] env[61970]: DEBUG nova.compute.manager [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1237.260776] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c201cd8f-89b8-4c2e-8a85-db39b681a970 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.265739] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1237.265906] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1237.266090] env[61970]: DEBUG nova.network.neutron [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1237.966288] env[61970]: DEBUG nova.network.neutron [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Updating instance_info_cache with network_info: [{"id": "be465191-5965-4079-a636-95bfbd16dd1d", "address": "fa:16:3e:b0:0a:ef", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe465191-59", "ovs_interfaceid": "be465191-5965-4079-a636-95bfbd16dd1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1238.469457] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1238.672193] env[61970]: DEBUG nova.compute.manager [req-1f8f3c26-3eeb-4420-82b6-26bb626ba670 req-0f421574-8465-445c-885f-d7122fd4dfda service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Received event network-vif-unplugged-be465191-5965-4079-a636-95bfbd16dd1d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1238.672423] env[61970]: DEBUG oslo_concurrency.lockutils [req-1f8f3c26-3eeb-4420-82b6-26bb626ba670 req-0f421574-8465-445c-885f-d7122fd4dfda service nova] Acquiring lock "bead8709-273f-433f-9d06-4c01510aaca1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.672630] env[61970]: DEBUG oslo_concurrency.lockutils [req-1f8f3c26-3eeb-4420-82b6-26bb626ba670 req-0f421574-8465-445c-885f-d7122fd4dfda service nova] Lock "bead8709-273f-433f-9d06-4c01510aaca1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.672801] env[61970]: DEBUG oslo_concurrency.lockutils [req-1f8f3c26-3eeb-4420-82b6-26bb626ba670 req-0f421574-8465-445c-885f-d7122fd4dfda service nova] Lock "bead8709-273f-433f-9d06-4c01510aaca1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1238.672970] env[61970]: DEBUG nova.compute.manager [req-1f8f3c26-3eeb-4420-82b6-26bb626ba670 req-0f421574-8465-445c-885f-d7122fd4dfda service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] No waiting events found dispatching network-vif-unplugged-be465191-5965-4079-a636-95bfbd16dd1d {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1238.673157] env[61970]: WARNING nova.compute.manager [req-1f8f3c26-3eeb-4420-82b6-26bb626ba670 req-0f421574-8465-445c-885f-d7122fd4dfda service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Received unexpected event network-vif-unplugged-be465191-5965-4079-a636-95bfbd16dd1d for instance with vm_state shelved and task_state shelving_offloading. [ 1238.759690] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1238.760600] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab313ec-ed0f-45c6-9280-dd1fdb5e98c1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.767917] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1238.768150] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-05012307-8a22-4fe1-8cda-fc076daeb277 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.836359] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1238.836513] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1238.836692] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Deleting the datastore file [datastore2] bead8709-273f-433f-9d06-4c01510aaca1 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1238.836956] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1fbaf440-2dc1-4500-8e45-b44cfe3cccf8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.843908] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1238.843908] env[61970]: value = "task-1356297" [ 1238.843908] env[61970]: _type = "Task" [ 1238.843908] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.851889] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356297, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.354401] env[61970]: DEBUG oslo_vmware.api [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356297, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135415} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.354931] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1239.354931] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1239.355061] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1239.378620] env[61970]: INFO nova.scheduler.client.report [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Deleted allocations for instance bead8709-273f-433f-9d06-4c01510aaca1 [ 1239.883311] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1239.883576] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1239.883799] env[61970]: DEBUG nova.objects.instance [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lazy-loading 'resources' on Instance uuid bead8709-273f-433f-9d06-4c01510aaca1 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1240.386764] env[61970]: DEBUG nova.objects.instance [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lazy-loading 'numa_topology' on Instance uuid bead8709-273f-433f-9d06-4c01510aaca1 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1240.700584] env[61970]: DEBUG nova.compute.manager [req-d8493dad-98ea-4236-84a3-3cde357ed327 req-b575d00b-3db8-4b1e-a056-939d75a60cb6 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Received event network-changed-be465191-5965-4079-a636-95bfbd16dd1d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1240.700820] env[61970]: DEBUG nova.compute.manager [req-d8493dad-98ea-4236-84a3-3cde357ed327 req-b575d00b-3db8-4b1e-a056-939d75a60cb6 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Refreshing instance network info cache due to event network-changed-be465191-5965-4079-a636-95bfbd16dd1d. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1240.701057] env[61970]: DEBUG oslo_concurrency.lockutils [req-d8493dad-98ea-4236-84a3-3cde357ed327 req-b575d00b-3db8-4b1e-a056-939d75a60cb6 service nova] Acquiring lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1240.701208] env[61970]: DEBUG oslo_concurrency.lockutils [req-d8493dad-98ea-4236-84a3-3cde357ed327 req-b575d00b-3db8-4b1e-a056-939d75a60cb6 service nova] Acquired lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1240.701372] env[61970]: DEBUG nova.network.neutron [req-d8493dad-98ea-4236-84a3-3cde357ed327 req-b575d00b-3db8-4b1e-a056-939d75a60cb6 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Refreshing network info cache for port be465191-5965-4079-a636-95bfbd16dd1d {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1240.889721] env[61970]: DEBUG nova.objects.base [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1240.915128] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-314175f4-c88a-435f-893c-a4b36fbfd818 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.922479] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76266e95-4b25-407b-8cf4-ef767d7a2d00 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.611780] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d15bba4d-3d15-4612-95f4-833f40fbd207 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.621170] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a56275c4-0e04-4f1f-8bd1-c03a797ef96e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.634402] env[61970]: DEBUG nova.compute.provider_tree [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1241.806301] env[61970]: DEBUG nova.network.neutron [req-d8493dad-98ea-4236-84a3-3cde357ed327 req-b575d00b-3db8-4b1e-a056-939d75a60cb6 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Updated VIF entry in instance network info cache for port be465191-5965-4079-a636-95bfbd16dd1d. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1241.806654] env[61970]: DEBUG nova.network.neutron [req-d8493dad-98ea-4236-84a3-3cde357ed327 req-b575d00b-3db8-4b1e-a056-939d75a60cb6 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Updating instance_info_cache with network_info: [{"id": "be465191-5965-4079-a636-95bfbd16dd1d", "address": "fa:16:3e:b0:0a:ef", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapbe465191-59", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1242.115646] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "bead8709-273f-433f-9d06-4c01510aaca1" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1242.137232] env[61970]: DEBUG nova.scheduler.client.report [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1242.309310] env[61970]: DEBUG oslo_concurrency.lockutils [req-d8493dad-98ea-4236-84a3-3cde357ed327 req-b575d00b-3db8-4b1e-a056-939d75a60cb6 service nova] Releasing lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1242.642357] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.759s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1243.150453] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ddd97b75-857c-42a8-8fe3-623d7c6b7d23 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "bead8709-273f-433f-9d06-4c01510aaca1" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.454s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1243.151349] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "bead8709-273f-433f-9d06-4c01510aaca1" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.036s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1243.151537] env[61970]: INFO nova.compute.manager [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Unshelving [ 1244.175118] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1244.175388] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1244.175603] env[61970]: DEBUG nova.objects.instance [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lazy-loading 'pci_requests' on Instance uuid bead8709-273f-433f-9d06-4c01510aaca1 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1244.680132] env[61970]: DEBUG nova.objects.instance [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lazy-loading 'numa_topology' on Instance uuid bead8709-273f-433f-9d06-4c01510aaca1 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1245.182887] env[61970]: INFO nova.compute.claims [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1246.217387] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f8c3af9-4975-4c7e-ac54-9eff683d78e8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.224967] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df3e7050-28ba-427b-aa09-8523d06a6390 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.253443] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a09f1d8-8e74-440c-bfd0-484da3843bb9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.259702] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed01f1e-b71a-4255-968b-859918ae38a8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.271878] env[61970]: DEBUG nova.compute.provider_tree [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1246.775506] env[61970]: DEBUG nova.scheduler.client.report [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1247.281223] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.106s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1247.307840] env[61970]: INFO nova.network.neutron [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Updating port be465191-5965-4079-a636-95bfbd16dd1d with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1248.695065] env[61970]: DEBUG nova.compute.manager [req-dc47b8b4-d351-42b8-9f0c-3d4bf1dd15c9 req-be04eb8b-cb64-46ce-9ed1-1a948dc612d9 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Received event network-vif-plugged-be465191-5965-4079-a636-95bfbd16dd1d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1248.695335] env[61970]: DEBUG oslo_concurrency.lockutils [req-dc47b8b4-d351-42b8-9f0c-3d4bf1dd15c9 req-be04eb8b-cb64-46ce-9ed1-1a948dc612d9 service nova] Acquiring lock "bead8709-273f-433f-9d06-4c01510aaca1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1248.695517] env[61970]: DEBUG oslo_concurrency.lockutils [req-dc47b8b4-d351-42b8-9f0c-3d4bf1dd15c9 req-be04eb8b-cb64-46ce-9ed1-1a948dc612d9 service nova] Lock "bead8709-273f-433f-9d06-4c01510aaca1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1248.695687] env[61970]: DEBUG oslo_concurrency.lockutils [req-dc47b8b4-d351-42b8-9f0c-3d4bf1dd15c9 req-be04eb8b-cb64-46ce-9ed1-1a948dc612d9 service nova] Lock "bead8709-273f-433f-9d06-4c01510aaca1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1248.695869] env[61970]: DEBUG nova.compute.manager [req-dc47b8b4-d351-42b8-9f0c-3d4bf1dd15c9 req-be04eb8b-cb64-46ce-9ed1-1a948dc612d9 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] No waiting events found dispatching network-vif-plugged-be465191-5965-4079-a636-95bfbd16dd1d {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1248.696039] env[61970]: WARNING nova.compute.manager [req-dc47b8b4-d351-42b8-9f0c-3d4bf1dd15c9 req-be04eb8b-cb64-46ce-9ed1-1a948dc612d9 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Received unexpected event network-vif-plugged-be465191-5965-4079-a636-95bfbd16dd1d for instance with vm_state shelved_offloaded and task_state spawning. [ 1248.781724] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1248.782063] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1248.782342] env[61970]: DEBUG nova.network.neutron [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1249.479272] env[61970]: DEBUG nova.network.neutron [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Updating instance_info_cache with network_info: [{"id": "be465191-5965-4079-a636-95bfbd16dd1d", "address": "fa:16:3e:b0:0a:ef", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe465191-59", "ovs_interfaceid": "be465191-5965-4079-a636-95bfbd16dd1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1249.981599] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1250.007381] env[61970]: DEBUG nova.virt.hardware [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='afe7f824a292afbed93c6c84644cc8e6',container_format='bare',created_at=2024-10-24T13:49:35Z,direct_url=,disk_format='vmdk',id=052216c7-6b4d-4f15-bb1f-25aa97395548,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-150880015-shelved',owner='09033ee688384ca287b4fdad6e67cb1f',properties=ImageMetaProps,protected=,size=31669248,status='active',tags=,updated_at=2024-10-24T13:49:47Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1250.007667] env[61970]: DEBUG nova.virt.hardware [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1250.007849] env[61970]: DEBUG nova.virt.hardware [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1250.008053] env[61970]: DEBUG nova.virt.hardware [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1250.008208] env[61970]: DEBUG nova.virt.hardware [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1250.008357] env[61970]: DEBUG nova.virt.hardware [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1250.008566] env[61970]: DEBUG nova.virt.hardware [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1250.008733] env[61970]: DEBUG nova.virt.hardware [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1250.008925] env[61970]: DEBUG nova.virt.hardware [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1250.009077] env[61970]: DEBUG nova.virt.hardware [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1250.009257] env[61970]: DEBUG nova.virt.hardware [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1250.010478] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c320242c-e281-439e-8a40-d4c1fd212b65 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.018528] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0078113c-806d-4c54-a4c7-b2987ddcbe77 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.031376] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:0a:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '298bb8ef-4765-494c-b157-7a349218bd1e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'be465191-5965-4079-a636-95bfbd16dd1d', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1250.038415] env[61970]: DEBUG oslo.service.loopingcall [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1250.038636] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1250.038834] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-539fcd25-947e-4382-a08b-1cc68c1b4171 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.056721] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1250.056721] env[61970]: value = "task-1356298" [ 1250.056721] env[61970]: _type = "Task" [ 1250.056721] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.063647] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356298, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.566691] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356298, 'name': CreateVM_Task, 'duration_secs': 0.291117} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.566882] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1250.567623] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/052216c7-6b4d-4f15-bb1f-25aa97395548" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1250.567810] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "[datastore2] devstack-image-cache_base/052216c7-6b4d-4f15-bb1f-25aa97395548" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1250.568209] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/052216c7-6b4d-4f15-bb1f-25aa97395548" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1250.568508] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cd43641-e4b6-49e6-9aef-0ad6b4c5fbcd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.572721] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1250.572721] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]527ceef1-2492-6c0d-c501-e8e47b3430a2" [ 1250.572721] env[61970]: _type = "Task" [ 1250.572721] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.579800] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]527ceef1-2492-6c0d-c501-e8e47b3430a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.721452] env[61970]: DEBUG nova.compute.manager [req-a624d8be-c9a0-43bb-939d-a582f1d3526a req-d6ddc430-3aac-438e-bc1b-aed93f64b901 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Received event network-changed-be465191-5965-4079-a636-95bfbd16dd1d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1250.721711] env[61970]: DEBUG nova.compute.manager [req-a624d8be-c9a0-43bb-939d-a582f1d3526a req-d6ddc430-3aac-438e-bc1b-aed93f64b901 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Refreshing instance network info cache due to event network-changed-be465191-5965-4079-a636-95bfbd16dd1d. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1250.721940] env[61970]: DEBUG oslo_concurrency.lockutils [req-a624d8be-c9a0-43bb-939d-a582f1d3526a req-d6ddc430-3aac-438e-bc1b-aed93f64b901 service nova] Acquiring lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1250.722149] env[61970]: DEBUG oslo_concurrency.lockutils [req-a624d8be-c9a0-43bb-939d-a582f1d3526a req-d6ddc430-3aac-438e-bc1b-aed93f64b901 service nova] Acquired lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1250.722267] env[61970]: DEBUG nova.network.neutron [req-a624d8be-c9a0-43bb-939d-a582f1d3526a req-d6ddc430-3aac-438e-bc1b-aed93f64b901 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Refreshing network info cache for port be465191-5965-4079-a636-95bfbd16dd1d {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1251.082277] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "[datastore2] devstack-image-cache_base/052216c7-6b4d-4f15-bb1f-25aa97395548" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1251.082630] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Processing image 052216c7-6b4d-4f15-bb1f-25aa97395548 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1251.082767] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/052216c7-6b4d-4f15-bb1f-25aa97395548/052216c7-6b4d-4f15-bb1f-25aa97395548.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1251.082921] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquired lock "[datastore2] devstack-image-cache_base/052216c7-6b4d-4f15-bb1f-25aa97395548/052216c7-6b4d-4f15-bb1f-25aa97395548.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1251.083120] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1251.083378] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ae205eb9-f48e-4e88-9a54-b66e439a7d8d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.100109] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1251.100282] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1251.100981] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c560b419-2848-47b7-a5ad-dffd6b88500b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.105723] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1251.105723] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526c6517-3389-ef8f-b9af-b0cdb60be48d" [ 1251.105723] env[61970]: _type = "Task" [ 1251.105723] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.112707] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526c6517-3389-ef8f-b9af-b0cdb60be48d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.450691] env[61970]: DEBUG nova.network.neutron [req-a624d8be-c9a0-43bb-939d-a582f1d3526a req-d6ddc430-3aac-438e-bc1b-aed93f64b901 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Updated VIF entry in instance network info cache for port be465191-5965-4079-a636-95bfbd16dd1d. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1251.451103] env[61970]: DEBUG nova.network.neutron [req-a624d8be-c9a0-43bb-939d-a582f1d3526a req-d6ddc430-3aac-438e-bc1b-aed93f64b901 service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Updating instance_info_cache with network_info: [{"id": "be465191-5965-4079-a636-95bfbd16dd1d", "address": "fa:16:3e:b0:0a:ef", "network": {"id": "a2a4345e-b03b-455f-b1c1-e85917fcf18d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1625033226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09033ee688384ca287b4fdad6e67cb1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe465191-59", "ovs_interfaceid": "be465191-5965-4079-a636-95bfbd16dd1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1251.615805] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Preparing fetch location {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1251.616013] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Fetch image to [datastore2] OSTACK_IMG_f29d72bc-8279-434b-aefb-62ae5afd2f94/OSTACK_IMG_f29d72bc-8279-434b-aefb-62ae5afd2f94.vmdk {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1251.616208] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Downloading stream optimized image 052216c7-6b4d-4f15-bb1f-25aa97395548 to [datastore2] OSTACK_IMG_f29d72bc-8279-434b-aefb-62ae5afd2f94/OSTACK_IMG_f29d72bc-8279-434b-aefb-62ae5afd2f94.vmdk on the data store datastore2 as vApp {{(pid=61970) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1251.616377] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Downloading image file data 052216c7-6b4d-4f15-bb1f-25aa97395548 to the ESX as VM named 'OSTACK_IMG_f29d72bc-8279-434b-aefb-62ae5afd2f94' {{(pid=61970) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1251.679112] env[61970]: DEBUG oslo_vmware.rw_handles [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1251.679112] env[61970]: value = "resgroup-9" [ 1251.679112] env[61970]: _type = "ResourcePool" [ 1251.679112] env[61970]: }. {{(pid=61970) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1251.679402] env[61970]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-23a0e2d9-b43c-404b-b64e-9b405f7eb0ea {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.698915] env[61970]: DEBUG oslo_vmware.rw_handles [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lease: (returnval){ [ 1251.698915] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526a05c1-b52c-0f15-0daf-d300e66b9b2d" [ 1251.698915] env[61970]: _type = "HttpNfcLease" [ 1251.698915] env[61970]: } obtained for vApp import into resource pool (val){ [ 1251.698915] env[61970]: value = "resgroup-9" [ 1251.698915] env[61970]: _type = "ResourcePool" [ 1251.698915] env[61970]: }. {{(pid=61970) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1251.699244] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the lease: (returnval){ [ 1251.699244] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526a05c1-b52c-0f15-0daf-d300e66b9b2d" [ 1251.699244] env[61970]: _type = "HttpNfcLease" [ 1251.699244] env[61970]: } to be ready. {{(pid=61970) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1251.705123] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1251.705123] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526a05c1-b52c-0f15-0daf-d300e66b9b2d" [ 1251.705123] env[61970]: _type = "HttpNfcLease" [ 1251.705123] env[61970]: } is initializing. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1251.954115] env[61970]: DEBUG oslo_concurrency.lockutils [req-a624d8be-c9a0-43bb-939d-a582f1d3526a req-d6ddc430-3aac-438e-bc1b-aed93f64b901 service nova] Releasing lock "refresh_cache-bead8709-273f-433f-9d06-4c01510aaca1" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1252.207531] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1252.207531] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526a05c1-b52c-0f15-0daf-d300e66b9b2d" [ 1252.207531] env[61970]: _type = "HttpNfcLease" [ 1252.207531] env[61970]: } is ready. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1252.208040] env[61970]: DEBUG oslo_vmware.rw_handles [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1252.208040] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526a05c1-b52c-0f15-0daf-d300e66b9b2d" [ 1252.208040] env[61970]: _type = "HttpNfcLease" [ 1252.208040] env[61970]: }. {{(pid=61970) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1252.208576] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1dae0c3-84c4-4048-8521-f98caa081163 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.215793] env[61970]: DEBUG oslo_vmware.rw_handles [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5233151f-e0da-ceef-e8b1-4a93165c9318/disk-0.vmdk from lease info. {{(pid=61970) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1252.216007] env[61970]: DEBUG oslo_vmware.rw_handles [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Creating HTTP connection to write to file with size = 31669248 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5233151f-e0da-ceef-e8b1-4a93165c9318/disk-0.vmdk. {{(pid=61970) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1252.278184] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-322728f7-fd2c-4b98-9712-e2834c4865ef {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.304155] env[61970]: DEBUG oslo_vmware.rw_handles [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Completed reading data from the image iterator. {{(pid=61970) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1253.304570] env[61970]: DEBUG oslo_vmware.rw_handles [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5233151f-e0da-ceef-e8b1-4a93165c9318/disk-0.vmdk. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1253.305689] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e6c314-edcf-4467-815f-c027ec3fb429 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.313265] env[61970]: DEBUG oslo_vmware.rw_handles [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5233151f-e0da-ceef-e8b1-4a93165c9318/disk-0.vmdk is in state: ready. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1253.313471] env[61970]: DEBUG oslo_vmware.rw_handles [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5233151f-e0da-ceef-e8b1-4a93165c9318/disk-0.vmdk. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1253.313768] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-95d64d39-799c-4cad-8971-9d42864862f5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.493636] env[61970]: DEBUG oslo_vmware.rw_handles [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5233151f-e0da-ceef-e8b1-4a93165c9318/disk-0.vmdk. {{(pid=61970) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1253.493861] env[61970]: INFO nova.virt.vmwareapi.images [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Downloaded image file data 052216c7-6b4d-4f15-bb1f-25aa97395548 [ 1253.494837] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea4a633-9f53-44e6-97b0-67c18988e529 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.509541] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b65ccaa4-7dff-4be6-acf0-0a03fdbb2048 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.538090] env[61970]: INFO nova.virt.vmwareapi.images [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] The imported VM was unregistered [ 1253.540238] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Caching image {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1253.540471] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Creating directory with path [datastore2] devstack-image-cache_base/052216c7-6b4d-4f15-bb1f-25aa97395548 {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1253.540781] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-abb4d235-33d4-443c-80b7-600ebb02d37c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.550260] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Created directory with path [datastore2] devstack-image-cache_base/052216c7-6b4d-4f15-bb1f-25aa97395548 {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1253.550442] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_f29d72bc-8279-434b-aefb-62ae5afd2f94/OSTACK_IMG_f29d72bc-8279-434b-aefb-62ae5afd2f94.vmdk to [datastore2] devstack-image-cache_base/052216c7-6b4d-4f15-bb1f-25aa97395548/052216c7-6b4d-4f15-bb1f-25aa97395548.vmdk. {{(pid=61970) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1253.550683] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-5ed96b2e-23cd-458c-981c-aeb69af49bfb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.556782] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1253.556782] env[61970]: value = "task-1356301" [ 1253.556782] env[61970]: _type = "Task" [ 1253.556782] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.563791] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356301, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.066837] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356301, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.568184] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356301, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.069075] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356301, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.570178] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356301, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.070930] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356301, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.225342} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.071162] env[61970]: INFO nova.virt.vmwareapi.ds_util [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_f29d72bc-8279-434b-aefb-62ae5afd2f94/OSTACK_IMG_f29d72bc-8279-434b-aefb-62ae5afd2f94.vmdk to [datastore2] devstack-image-cache_base/052216c7-6b4d-4f15-bb1f-25aa97395548/052216c7-6b4d-4f15-bb1f-25aa97395548.vmdk. [ 1256.071361] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Cleaning up location [datastore2] OSTACK_IMG_f29d72bc-8279-434b-aefb-62ae5afd2f94 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1256.071530] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_f29d72bc-8279-434b-aefb-62ae5afd2f94 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1256.071813] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ea057279-4c8c-4a62-a292-6e54e6939dfb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.078073] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1256.078073] env[61970]: value = "task-1356302" [ 1256.078073] env[61970]: _type = "Task" [ 1256.078073] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.085055] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356302, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.588708] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356302, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.033371} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.589073] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1256.589133] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Releasing lock "[datastore2] devstack-image-cache_base/052216c7-6b4d-4f15-bb1f-25aa97395548/052216c7-6b4d-4f15-bb1f-25aa97395548.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1256.589453] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/052216c7-6b4d-4f15-bb1f-25aa97395548/052216c7-6b4d-4f15-bb1f-25aa97395548.vmdk to [datastore2] bead8709-273f-433f-9d06-4c01510aaca1/bead8709-273f-433f-9d06-4c01510aaca1.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1256.589744] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8bc96f0c-065e-4eb9-913a-472c0fa6bbba {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.595962] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1256.595962] env[61970]: value = "task-1356303" [ 1256.595962] env[61970]: _type = "Task" [ 1256.595962] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.604192] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356303, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.105750] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356303, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.608024] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356303, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.109043] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356303, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.609842] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356303, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.110711] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356303, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.185939} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.110989] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/052216c7-6b4d-4f15-bb1f-25aa97395548/052216c7-6b4d-4f15-bb1f-25aa97395548.vmdk to [datastore2] bead8709-273f-433f-9d06-4c01510aaca1/bead8709-273f-433f-9d06-4c01510aaca1.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1259.111758] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee33cb2c-0c50-442f-b907-60d6103943b8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.132511] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] bead8709-273f-433f-9d06-4c01510aaca1/bead8709-273f-433f-9d06-4c01510aaca1.vmdk or device None with type streamOptimized {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1259.132766] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c79ddb84-fc7e-47a4-a2b4-a96c3f58594a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.152988] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1259.152988] env[61970]: value = "task-1356304" [ 1259.152988] env[61970]: _type = "Task" [ 1259.152988] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.160498] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356304, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.666278] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356304, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.163427] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356304, 'name': ReconfigVM_Task, 'duration_secs': 0.656835} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.163784] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Reconfigured VM instance instance-0000006b to attach disk [datastore2] bead8709-273f-433f-9d06-4c01510aaca1/bead8709-273f-433f-9d06-4c01510aaca1.vmdk or device None with type streamOptimized {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1260.164375] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-86e95d8f-be76-494c-8749-cecc8e23bf92 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.172634] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1260.172634] env[61970]: value = "task-1356305" [ 1260.172634] env[61970]: _type = "Task" [ 1260.172634] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.177783] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356305, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.680299] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356305, 'name': Rename_Task, 'duration_secs': 0.136394} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.680688] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1260.680778] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-039818cb-2b45-4359-860c-a39b63152109 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.686805] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1260.686805] env[61970]: value = "task-1356306" [ 1260.686805] env[61970]: _type = "Task" [ 1260.686805] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.693848] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356306, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.197235] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356306, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.696702] env[61970]: DEBUG oslo_vmware.api [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356306, 'name': PowerOnVM_Task, 'duration_secs': 0.838798} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.697151] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1261.785173] env[61970]: DEBUG nova.compute.manager [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1261.786075] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95899761-cf72-4eea-87fa-19d933b279e5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.303051] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7ae89c12-b9b8-4cc4-8e7b-b5dc00aedeef tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "bead8709-273f-433f-9d06-4c01510aaca1" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.151s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1263.182732] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "bead8709-273f-433f-9d06-4c01510aaca1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1263.183113] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "bead8709-273f-433f-9d06-4c01510aaca1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1263.183284] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "bead8709-273f-433f-9d06-4c01510aaca1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1263.183482] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "bead8709-273f-433f-9d06-4c01510aaca1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1263.183661] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "bead8709-273f-433f-9d06-4c01510aaca1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1263.185787] env[61970]: INFO nova.compute.manager [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Terminating instance [ 1263.187435] env[61970]: DEBUG nova.compute.manager [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1263.187655] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1263.188492] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f9c3c3-e925-4cf9-8bf6-bf314e32df41 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.197808] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1263.198050] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e2bc9001-99d3-49ff-a44b-2393a02d61ce {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.204474] env[61970]: DEBUG oslo_vmware.api [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1263.204474] env[61970]: value = "task-1356307" [ 1263.204474] env[61970]: _type = "Task" [ 1263.204474] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.212376] env[61970]: DEBUG oslo_vmware.api [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356307, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.714911] env[61970]: DEBUG oslo_vmware.api [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356307, 'name': PowerOffVM_Task, 'duration_secs': 0.174573} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.716593] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1263.716593] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1263.716593] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-476fe9ca-8582-4adf-9988-c2de1c179686 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.777156] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1263.777156] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1263.777156] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Deleting the datastore file [datastore2] bead8709-273f-433f-9d06-4c01510aaca1 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1263.777156] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d3fee34a-8cd6-4f01-ae74-302c8165daec {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.783058] env[61970]: DEBUG oslo_vmware.api [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for the task: (returnval){ [ 1263.783058] env[61970]: value = "task-1356309" [ 1263.783058] env[61970]: _type = "Task" [ 1263.783058] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.790518] env[61970]: DEBUG oslo_vmware.api [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356309, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.293388] env[61970]: DEBUG oslo_vmware.api [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Task: {'id': task-1356309, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122849} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.293754] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1264.293892] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1264.293984] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1264.294188] env[61970]: INFO nova.compute.manager [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1264.294430] env[61970]: DEBUG oslo.service.loopingcall [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1264.294626] env[61970]: DEBUG nova.compute.manager [-] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1264.294714] env[61970]: DEBUG nova.network.neutron [-] [instance: bead8709-273f-433f-9d06-4c01510aaca1] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1264.707414] env[61970]: DEBUG nova.compute.manager [req-afd1d42e-9ea0-46e4-b9e1-8e5d972558c2 req-98304011-1516-4cdf-8e2f-964fbe3f0ada service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Received event network-vif-deleted-be465191-5965-4079-a636-95bfbd16dd1d {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1264.707414] env[61970]: INFO nova.compute.manager [req-afd1d42e-9ea0-46e4-b9e1-8e5d972558c2 req-98304011-1516-4cdf-8e2f-964fbe3f0ada service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Neutron deleted interface be465191-5965-4079-a636-95bfbd16dd1d; detaching it from the instance and deleting it from the info cache [ 1264.707414] env[61970]: DEBUG nova.network.neutron [req-afd1d42e-9ea0-46e4-b9e1-8e5d972558c2 req-98304011-1516-4cdf-8e2f-964fbe3f0ada service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1265.185906] env[61970]: DEBUG nova.network.neutron [-] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1265.209367] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-83f21ab0-1095-476f-9e3e-0db43c580709 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.219138] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ae76a2-5dcf-465c-8e51-5f551535180a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.240472] env[61970]: DEBUG nova.compute.manager [req-afd1d42e-9ea0-46e4-b9e1-8e5d972558c2 req-98304011-1516-4cdf-8e2f-964fbe3f0ada service nova] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Detach interface failed, port_id=be465191-5965-4079-a636-95bfbd16dd1d, reason: Instance bead8709-273f-433f-9d06-4c01510aaca1 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1265.259210] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1265.688723] env[61970]: INFO nova.compute.manager [-] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Took 1.39 seconds to deallocate network for instance. [ 1266.195966] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1266.196119] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1266.196349] env[61970]: DEBUG nova.objects.instance [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lazy-loading 'resources' on Instance uuid bead8709-273f-433f-9d06-4c01510aaca1 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1266.730382] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e87199f-af84-47b3-8781-06483ce10e97 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.737957] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e2c6562-3f0b-4fc1-8811-4c5219a29eb7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.766786] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bc7e1fd-ee70-47f3-952b-fbe81ff2de65 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.773441] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e667f5d-a3ab-4e4e-a0c3-b4ac38886fa7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.785956] env[61970]: DEBUG nova.compute.provider_tree [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1267.260237] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1267.289367] env[61970]: DEBUG nova.scheduler.client.report [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1267.794675] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.598s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1267.815565] env[61970]: INFO nova.scheduler.client.report [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Deleted allocations for instance bead8709-273f-433f-9d06-4c01510aaca1 [ 1268.323794] env[61970]: DEBUG oslo_concurrency.lockutils [None req-6d0d3afd-debf-42a5-830d-67c3fc994dd8 tempest-ServerActionsTestOtherB-911101615 tempest-ServerActionsTestOtherB-911101615-project-member] Lock "bead8709-273f-433f-9d06-4c01510aaca1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.141s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1270.255621] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1270.259212] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1271.259607] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1271.259878] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1271.260037] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61970) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1271.260199] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1271.763665] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1271.763921] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1271.764100] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1271.764256] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61970) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1271.765653] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb8b1822-1efd-48c7-85e4-03fe777296f0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.773847] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d03f7ed-0e81-477a-b6ce-75590e6fca59 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.787628] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9ee065-52a3-4255-89aa-ca17d7cc5e40 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.793891] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-051832ec-812a-4088-8ad6-5db75fdff8b9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.821876] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181204MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61970) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1271.822046] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1271.822300] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1272.847751] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1272.847990] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1272.866797] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33bf47b7-7db1-4ebb-9539-2e8d32b0cfde {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.872988] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04525eb6-30ca-4620-b787-85afdb8acdbd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.907131] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e75f29a3-9669-4146-8991-bfb2ed8c76b7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.914621] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc393ea1-7b13-4d27-a718-49246d4f86a0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.927602] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1273.430718] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1273.936361] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1273.936713] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.114s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1274.937187] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1274.937479] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Starting heal instance info cache {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1274.937479] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Rebuilding the list of instances to heal {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1275.441198] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Didn't find any instances for network info cache update. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1275.441456] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1276.089032] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "d468d634-e0ae-4407-beeb-4c56049ba203" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1276.089328] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "d468d634-e0ae-4407-beeb-4c56049ba203" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1276.396089] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "9d77a294-09ca-447c-924b-4feb0124dc70" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1276.396392] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "9d77a294-09ca-447c-924b-4feb0124dc70" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1276.591369] env[61970]: DEBUG nova.compute.manager [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1276.899511] env[61970]: DEBUG nova.compute.manager [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1277.113754] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1277.114039] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1277.115496] env[61970]: INFO nova.compute.claims [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1277.418129] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1278.159601] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03fe075-1098-44c1-a3dc-960b7ed687ee {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.166904] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9b7720-a146-4b9b-80b7-967f94f475e0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.195600] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c3aef01-be50-4b9f-a4b7-665083e2a113 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.202928] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f87339b3-363e-4d18-9237-8e7adc0f99b6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.216952] env[61970]: DEBUG nova.compute.provider_tree [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1278.719731] env[61970]: DEBUG nova.scheduler.client.report [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1279.225184] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.111s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1279.225694] env[61970]: DEBUG nova.compute.manager [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1279.228684] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.811s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1279.230078] env[61970]: INFO nova.compute.claims [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1279.733913] env[61970]: DEBUG nova.compute.utils [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1279.737690] env[61970]: DEBUG nova.compute.manager [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1279.737891] env[61970]: DEBUG nova.network.neutron [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1279.775162] env[61970]: DEBUG nova.policy [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '248175c9d03f483eadaadab6bec2a3c9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b739147231ef473288ef6c78eed1caf3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 1280.005878] env[61970]: DEBUG nova.network.neutron [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Successfully created port: 1a5feac9-a368-4804-bb4a-648e682d6682 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1280.239197] env[61970]: DEBUG nova.compute.manager [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1280.280272] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e98ba38-94e5-419b-bd81-1c5557b75dcb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.288442] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b266567-d0b1-4901-ab52-5baf21163c81 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.320543] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f50d73-06fb-4d66-a4d8-8ccd83a10320 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.327205] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817fce8b-677d-4586-b56d-1145d61bc573 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.340797] env[61970]: DEBUG nova.compute.provider_tree [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1280.843674] env[61970]: DEBUG nova.scheduler.client.report [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1281.251758] env[61970]: DEBUG nova.compute.manager [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1281.275502] env[61970]: DEBUG nova.virt.hardware [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1281.275751] env[61970]: DEBUG nova.virt.hardware [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1281.275913] env[61970]: DEBUG nova.virt.hardware [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1281.276112] env[61970]: DEBUG nova.virt.hardware [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1281.276264] env[61970]: DEBUG nova.virt.hardware [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1281.276414] env[61970]: DEBUG nova.virt.hardware [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1281.276619] env[61970]: DEBUG nova.virt.hardware [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1281.276779] env[61970]: DEBUG nova.virt.hardware [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1281.276948] env[61970]: DEBUG nova.virt.hardware [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1281.277129] env[61970]: DEBUG nova.virt.hardware [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1281.277306] env[61970]: DEBUG nova.virt.hardware [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1281.278196] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3323065-1762-4f76-b2f9-6b405bbf4c05 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.285756] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb92858-9c92-4b02-9ebd-f130d284c63b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.348412] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.120s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1281.348902] env[61970]: DEBUG nova.compute.manager [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1281.380293] env[61970]: DEBUG nova.compute.manager [req-bb723d80-ee72-415d-a616-44c9169baeb9 req-ff7edb24-e1dd-4cde-8566-962d962c59d6 service nova] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Received event network-vif-plugged-1a5feac9-a368-4804-bb4a-648e682d6682 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1281.380560] env[61970]: DEBUG oslo_concurrency.lockutils [req-bb723d80-ee72-415d-a616-44c9169baeb9 req-ff7edb24-e1dd-4cde-8566-962d962c59d6 service nova] Acquiring lock "d468d634-e0ae-4407-beeb-4c56049ba203-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1281.380782] env[61970]: DEBUG oslo_concurrency.lockutils [req-bb723d80-ee72-415d-a616-44c9169baeb9 req-ff7edb24-e1dd-4cde-8566-962d962c59d6 service nova] Lock "d468d634-e0ae-4407-beeb-4c56049ba203-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1281.381034] env[61970]: DEBUG oslo_concurrency.lockutils [req-bb723d80-ee72-415d-a616-44c9169baeb9 req-ff7edb24-e1dd-4cde-8566-962d962c59d6 service nova] Lock "d468d634-e0ae-4407-beeb-4c56049ba203-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1281.381469] env[61970]: DEBUG nova.compute.manager [req-bb723d80-ee72-415d-a616-44c9169baeb9 req-ff7edb24-e1dd-4cde-8566-962d962c59d6 service nova] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] No waiting events found dispatching network-vif-plugged-1a5feac9-a368-4804-bb4a-648e682d6682 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1281.381673] env[61970]: WARNING nova.compute.manager [req-bb723d80-ee72-415d-a616-44c9169baeb9 req-ff7edb24-e1dd-4cde-8566-962d962c59d6 service nova] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Received unexpected event network-vif-plugged-1a5feac9-a368-4804-bb4a-648e682d6682 for instance with vm_state building and task_state spawning. [ 1281.463849] env[61970]: DEBUG nova.network.neutron [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Successfully updated port: 1a5feac9-a368-4804-bb4a-648e682d6682 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1281.853496] env[61970]: DEBUG nova.compute.utils [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1281.854978] env[61970]: DEBUG nova.compute.manager [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1281.855619] env[61970]: DEBUG nova.network.neutron [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1281.893082] env[61970]: DEBUG nova.policy [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '248175c9d03f483eadaadab6bec2a3c9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b739147231ef473288ef6c78eed1caf3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 1281.966328] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "refresh_cache-d468d634-e0ae-4407-beeb-4c56049ba203" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1281.966508] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired lock "refresh_cache-d468d634-e0ae-4407-beeb-4c56049ba203" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1281.966624] env[61970]: DEBUG nova.network.neutron [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1282.118355] env[61970]: DEBUG nova.network.neutron [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Successfully created port: 1e976667-f251-46e2-ae84-6b74c304cf63 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1282.359011] env[61970]: DEBUG nova.compute.manager [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1282.503164] env[61970]: DEBUG nova.network.neutron [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1282.624055] env[61970]: DEBUG nova.network.neutron [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Updating instance_info_cache with network_info: [{"id": "1a5feac9-a368-4804-bb4a-648e682d6682", "address": "fa:16:3e:62:94:3e", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a5feac9-a3", "ovs_interfaceid": "1a5feac9-a368-4804-bb4a-648e682d6682", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1283.127149] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Releasing lock "refresh_cache-d468d634-e0ae-4407-beeb-4c56049ba203" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1283.127448] env[61970]: DEBUG nova.compute.manager [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Instance network_info: |[{"id": "1a5feac9-a368-4804-bb4a-648e682d6682", "address": "fa:16:3e:62:94:3e", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a5feac9-a3", "ovs_interfaceid": "1a5feac9-a368-4804-bb4a-648e682d6682", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1283.127912] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:94:3e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '669e4919-e0ad-4e23-9f23-4c5f2be0d858', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1a5feac9-a368-4804-bb4a-648e682d6682', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1283.135360] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Creating folder: Project (b739147231ef473288ef6c78eed1caf3). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1283.135641] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ddba65bb-0e3a-480e-9c2e-06a3f5176df3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.149148] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Created folder: Project (b739147231ef473288ef6c78eed1caf3) in parent group-v288740. [ 1283.149336] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Creating folder: Instances. Parent ref: group-v288899. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1283.149562] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53a8e6ef-3ca1-460b-ba26-3e6bdf25bc76 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.158283] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Created folder: Instances in parent group-v288899. [ 1283.158506] env[61970]: DEBUG oslo.service.loopingcall [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1283.158686] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1283.158875] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1c83326-a513-4cd6-a34c-5d30b4d84f8e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.177391] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1283.177391] env[61970]: value = "task-1356313" [ 1283.177391] env[61970]: _type = "Task" [ 1283.177391] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.184312] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356313, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1283.371394] env[61970]: DEBUG nova.compute.manager [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1283.400019] env[61970]: DEBUG nova.virt.hardware [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1283.400340] env[61970]: DEBUG nova.virt.hardware [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1283.400512] env[61970]: DEBUG nova.virt.hardware [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1283.400702] env[61970]: DEBUG nova.virt.hardware [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1283.400895] env[61970]: DEBUG nova.virt.hardware [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1283.401064] env[61970]: DEBUG nova.virt.hardware [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1283.401301] env[61970]: DEBUG nova.virt.hardware [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1283.401469] env[61970]: DEBUG nova.virt.hardware [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1283.401639] env[61970]: DEBUG nova.virt.hardware [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1283.401836] env[61970]: DEBUG nova.virt.hardware [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1283.402085] env[61970]: DEBUG nova.virt.hardware [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1283.402975] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5229a383-5eea-4f82-8623-ddf74c693e2e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.408197] env[61970]: DEBUG nova.compute.manager [req-3cd62de9-c757-4b3d-b565-24593c39f099 req-3e461957-ebda-45c9-adb6-caf6312fa77b service nova] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Received event network-changed-1a5feac9-a368-4804-bb4a-648e682d6682 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1283.408371] env[61970]: DEBUG nova.compute.manager [req-3cd62de9-c757-4b3d-b565-24593c39f099 req-3e461957-ebda-45c9-adb6-caf6312fa77b service nova] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Refreshing instance network info cache due to event network-changed-1a5feac9-a368-4804-bb4a-648e682d6682. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1283.408591] env[61970]: DEBUG oslo_concurrency.lockutils [req-3cd62de9-c757-4b3d-b565-24593c39f099 req-3e461957-ebda-45c9-adb6-caf6312fa77b service nova] Acquiring lock "refresh_cache-d468d634-e0ae-4407-beeb-4c56049ba203" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1283.408736] env[61970]: DEBUG oslo_concurrency.lockutils [req-3cd62de9-c757-4b3d-b565-24593c39f099 req-3e461957-ebda-45c9-adb6-caf6312fa77b service nova] Acquired lock "refresh_cache-d468d634-e0ae-4407-beeb-4c56049ba203" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1283.408897] env[61970]: DEBUG nova.network.neutron [req-3cd62de9-c757-4b3d-b565-24593c39f099 req-3e461957-ebda-45c9-adb6-caf6312fa77b service nova] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Refreshing network info cache for port 1a5feac9-a368-4804-bb4a-648e682d6682 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1283.414930] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820ad61e-3adc-43bc-b399-15f78b6006c1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.593642] env[61970]: DEBUG nova.network.neutron [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Successfully updated port: 1e976667-f251-46e2-ae84-6b74c304cf63 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1283.688051] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356313, 'name': CreateVM_Task, 'duration_secs': 0.304877} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1283.688051] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1283.688051] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1283.688296] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1283.688508] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1283.688756] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa24733f-5e6c-49e5-9855-dab1c2ce84e3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.692914] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1283.692914] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526d4057-a555-bae7-b59e-9bdb3f4d7f23" [ 1283.692914] env[61970]: _type = "Task" [ 1283.692914] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.700074] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526d4057-a555-bae7-b59e-9bdb3f4d7f23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.092929] env[61970]: DEBUG nova.network.neutron [req-3cd62de9-c757-4b3d-b565-24593c39f099 req-3e461957-ebda-45c9-adb6-caf6312fa77b service nova] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Updated VIF entry in instance network info cache for port 1a5feac9-a368-4804-bb4a-648e682d6682. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1284.093323] env[61970]: DEBUG nova.network.neutron [req-3cd62de9-c757-4b3d-b565-24593c39f099 req-3e461957-ebda-45c9-adb6-caf6312fa77b service nova] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Updating instance_info_cache with network_info: [{"id": "1a5feac9-a368-4804-bb4a-648e682d6682", "address": "fa:16:3e:62:94:3e", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a5feac9-a3", "ovs_interfaceid": "1a5feac9-a368-4804-bb4a-648e682d6682", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1284.096320] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "refresh_cache-9d77a294-09ca-447c-924b-4feb0124dc70" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1284.096453] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired lock "refresh_cache-9d77a294-09ca-447c-924b-4feb0124dc70" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1284.096606] env[61970]: DEBUG nova.network.neutron [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1284.204358] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]526d4057-a555-bae7-b59e-9bdb3f4d7f23, 'name': SearchDatastore_Task, 'duration_secs': 0.011146} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1284.204660] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1284.204902] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1284.205159] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1284.205314] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1284.205494] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1284.205762] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6b79f1e2-4902-40b9-933f-70b28a3dce8e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.214204] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1284.214377] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1284.215078] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12ae5d45-b012-41be-925f-c6aed827d80d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.219818] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1284.219818] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a844fb-234d-63e1-de94-ec99db077c6b" [ 1284.219818] env[61970]: _type = "Task" [ 1284.219818] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1284.227381] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a844fb-234d-63e1-de94-ec99db077c6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.595951] env[61970]: DEBUG oslo_concurrency.lockutils [req-3cd62de9-c757-4b3d-b565-24593c39f099 req-3e461957-ebda-45c9-adb6-caf6312fa77b service nova] Releasing lock "refresh_cache-d468d634-e0ae-4407-beeb-4c56049ba203" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1284.626954] env[61970]: DEBUG nova.network.neutron [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1284.730786] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52a844fb-234d-63e1-de94-ec99db077c6b, 'name': SearchDatastore_Task, 'duration_secs': 0.00836} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1284.731559] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40f5636b-9894-44f9-bf96-9d6adc568023 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.736397] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1284.736397] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]521ea608-aeb2-994c-f455-22cbffad9d10" [ 1284.736397] env[61970]: _type = "Task" [ 1284.736397] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1284.743554] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]521ea608-aeb2-994c-f455-22cbffad9d10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.744368] env[61970]: DEBUG nova.network.neutron [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Updating instance_info_cache with network_info: [{"id": "1e976667-f251-46e2-ae84-6b74c304cf63", "address": "fa:16:3e:cb:7d:1b", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e976667-f2", "ovs_interfaceid": "1e976667-f251-46e2-ae84-6b74c304cf63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1285.247233] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Releasing lock "refresh_cache-9d77a294-09ca-447c-924b-4feb0124dc70" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1285.247527] env[61970]: DEBUG nova.compute.manager [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Instance network_info: |[{"id": "1e976667-f251-46e2-ae84-6b74c304cf63", "address": "fa:16:3e:cb:7d:1b", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e976667-f2", "ovs_interfaceid": "1e976667-f251-46e2-ae84-6b74c304cf63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1285.247842] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]521ea608-aeb2-994c-f455-22cbffad9d10, 'name': SearchDatastore_Task, 'duration_secs': 0.008183} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.248192] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:7d:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '669e4919-e0ad-4e23-9f23-4c5f2be0d858', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1e976667-f251-46e2-ae84-6b74c304cf63', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1285.255483] env[61970]: DEBUG oslo.service.loopingcall [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1285.255632] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1285.255862] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] d468d634-e0ae-4407-beeb-4c56049ba203/d468d634-e0ae-4407-beeb-4c56049ba203.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1285.256119] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1285.256348] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c76947af-f962-45c7-a5ef-4aa44de6bae3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.258318] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1cdf5f14-2741-4d9a-992b-6d0a9994fdc7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.276947] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1285.276947] env[61970]: value = "task-1356315" [ 1285.276947] env[61970]: _type = "Task" [ 1285.276947] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.280741] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1285.280741] env[61970]: value = "task-1356314" [ 1285.280741] env[61970]: _type = "Task" [ 1285.280741] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.286262] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356315, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.290541] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356314, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.438523] env[61970]: DEBUG nova.compute.manager [req-d5277402-1a43-44ca-9df3-5f6f410a6eec req-ef213243-0008-4765-a811-d40ff3029849 service nova] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Received event network-vif-plugged-1e976667-f251-46e2-ae84-6b74c304cf63 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1285.438805] env[61970]: DEBUG oslo_concurrency.lockutils [req-d5277402-1a43-44ca-9df3-5f6f410a6eec req-ef213243-0008-4765-a811-d40ff3029849 service nova] Acquiring lock "9d77a294-09ca-447c-924b-4feb0124dc70-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1285.438973] env[61970]: DEBUG oslo_concurrency.lockutils [req-d5277402-1a43-44ca-9df3-5f6f410a6eec req-ef213243-0008-4765-a811-d40ff3029849 service nova] Lock "9d77a294-09ca-447c-924b-4feb0124dc70-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1285.439183] env[61970]: DEBUG oslo_concurrency.lockutils [req-d5277402-1a43-44ca-9df3-5f6f410a6eec req-ef213243-0008-4765-a811-d40ff3029849 service nova] Lock "9d77a294-09ca-447c-924b-4feb0124dc70-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1285.439402] env[61970]: DEBUG nova.compute.manager [req-d5277402-1a43-44ca-9df3-5f6f410a6eec req-ef213243-0008-4765-a811-d40ff3029849 service nova] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] No waiting events found dispatching network-vif-plugged-1e976667-f251-46e2-ae84-6b74c304cf63 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1285.439580] env[61970]: WARNING nova.compute.manager [req-d5277402-1a43-44ca-9df3-5f6f410a6eec req-ef213243-0008-4765-a811-d40ff3029849 service nova] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Received unexpected event network-vif-plugged-1e976667-f251-46e2-ae84-6b74c304cf63 for instance with vm_state building and task_state spawning. [ 1285.439813] env[61970]: DEBUG nova.compute.manager [req-d5277402-1a43-44ca-9df3-5f6f410a6eec req-ef213243-0008-4765-a811-d40ff3029849 service nova] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Received event network-changed-1e976667-f251-46e2-ae84-6b74c304cf63 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1285.440074] env[61970]: DEBUG nova.compute.manager [req-d5277402-1a43-44ca-9df3-5f6f410a6eec req-ef213243-0008-4765-a811-d40ff3029849 service nova] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Refreshing instance network info cache due to event network-changed-1e976667-f251-46e2-ae84-6b74c304cf63. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1285.440298] env[61970]: DEBUG oslo_concurrency.lockutils [req-d5277402-1a43-44ca-9df3-5f6f410a6eec req-ef213243-0008-4765-a811-d40ff3029849 service nova] Acquiring lock "refresh_cache-9d77a294-09ca-447c-924b-4feb0124dc70" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1285.440504] env[61970]: DEBUG oslo_concurrency.lockutils [req-d5277402-1a43-44ca-9df3-5f6f410a6eec req-ef213243-0008-4765-a811-d40ff3029849 service nova] Acquired lock "refresh_cache-9d77a294-09ca-447c-924b-4feb0124dc70" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1285.440726] env[61970]: DEBUG nova.network.neutron [req-d5277402-1a43-44ca-9df3-5f6f410a6eec req-ef213243-0008-4765-a811-d40ff3029849 service nova] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Refreshing network info cache for port 1e976667-f251-46e2-ae84-6b74c304cf63 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1285.788101] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356315, 'name': CreateVM_Task, 'duration_secs': 0.414755} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.788526] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1285.789211] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1285.789377] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1285.789697] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1285.792695] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e56b33a7-ac23-4991-a419-f85e8d0c5c77 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.794091] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356314, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.459866} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.794326] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] d468d634-e0ae-4407-beeb-4c56049ba203/d468d634-e0ae-4407-beeb-4c56049ba203.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1285.794539] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1285.795060] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0f6c5c82-6f41-4289-a1eb-c97495d2a6b4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.797953] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1285.797953] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b71c3c-886d-b437-c84e-6ed4547af647" [ 1285.797953] env[61970]: _type = "Task" [ 1285.797953] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.802196] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1285.802196] env[61970]: value = "task-1356316" [ 1285.802196] env[61970]: _type = "Task" [ 1285.802196] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.807327] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b71c3c-886d-b437-c84e-6ed4547af647, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.811504] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356316, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.120588] env[61970]: DEBUG nova.network.neutron [req-d5277402-1a43-44ca-9df3-5f6f410a6eec req-ef213243-0008-4765-a811-d40ff3029849 service nova] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Updated VIF entry in instance network info cache for port 1e976667-f251-46e2-ae84-6b74c304cf63. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1286.120964] env[61970]: DEBUG nova.network.neutron [req-d5277402-1a43-44ca-9df3-5f6f410a6eec req-ef213243-0008-4765-a811-d40ff3029849 service nova] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Updating instance_info_cache with network_info: [{"id": "1e976667-f251-46e2-ae84-6b74c304cf63", "address": "fa:16:3e:cb:7d:1b", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e976667-f2", "ovs_interfaceid": "1e976667-f251-46e2-ae84-6b74c304cf63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1286.310560] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356316, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.313874] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52b71c3c-886d-b437-c84e-6ed4547af647, 'name': SearchDatastore_Task, 'duration_secs': 0.008434} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.314165] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1286.314399] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1286.314628] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1286.314776] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1286.314953] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1286.315217] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e1e98565-9a24-43e3-9b74-461cf7fef209 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.322654] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1286.322842] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1286.323523] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6231eeb7-7175-47ae-aab4-8802642d8caf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.327766] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1286.327766] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52da5922-951f-2476-d3c0-d72179e7e053" [ 1286.327766] env[61970]: _type = "Task" [ 1286.327766] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.335821] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52da5922-951f-2476-d3c0-d72179e7e053, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.623785] env[61970]: DEBUG oslo_concurrency.lockutils [req-d5277402-1a43-44ca-9df3-5f6f410a6eec req-ef213243-0008-4765-a811-d40ff3029849 service nova] Releasing lock "refresh_cache-9d77a294-09ca-447c-924b-4feb0124dc70" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1286.812620] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356316, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.910198} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.813032] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1286.813630] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a628c05f-f105-4ef9-a7c2-dbcbe5b73275 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.834884] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] d468d634-e0ae-4407-beeb-4c56049ba203/d468d634-e0ae-4407-beeb-4c56049ba203.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1286.835143] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a034c682-833d-4de9-9200-a77624bcad47 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.856103] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52da5922-951f-2476-d3c0-d72179e7e053, 'name': SearchDatastore_Task, 'duration_secs': 0.00832} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.857707] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1286.857707] env[61970]: value = "task-1356317" [ 1286.857707] env[61970]: _type = "Task" [ 1286.857707] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.857916] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea2fcb11-77f7-498e-995b-e6a43fc5663b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.865088] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1286.865088] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e266bc-70e1-4117-58f0-13e081de89d4" [ 1286.865088] env[61970]: _type = "Task" [ 1286.865088] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.868069] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356317, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.874812] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e266bc-70e1-4117-58f0-13e081de89d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.369549] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356317, 'name': ReconfigVM_Task, 'duration_secs': 0.258084} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.372537] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Reconfigured VM instance instance-0000006c to attach disk [datastore2] d468d634-e0ae-4407-beeb-4c56049ba203/d468d634-e0ae-4407-beeb-4c56049ba203.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1287.373166] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a948c6fb-87f4-48d3-9a8b-e34bd88c65b8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.379678] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e266bc-70e1-4117-58f0-13e081de89d4, 'name': SearchDatastore_Task, 'duration_secs': 0.008846} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.380772] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1287.381051] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 9d77a294-09ca-447c-924b-4feb0124dc70/9d77a294-09ca-447c-924b-4feb0124dc70.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1287.381362] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1287.381362] env[61970]: value = "task-1356318" [ 1287.381362] env[61970]: _type = "Task" [ 1287.381362] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.381544] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-71ef994c-ccb6-4c69-8751-ee89e02c80be {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.390816] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356318, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.391905] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1287.391905] env[61970]: value = "task-1356319" [ 1287.391905] env[61970]: _type = "Task" [ 1287.391905] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.398738] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356319, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.892603] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356318, 'name': Rename_Task, 'duration_secs': 0.140086} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.895701] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1287.895946] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c8dee6e3-1ef3-40d3-8c35-53697109f0a4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.901635] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356319, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448576} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.902682] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 9d77a294-09ca-447c-924b-4feb0124dc70/9d77a294-09ca-447c-924b-4feb0124dc70.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1287.902894] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1287.903225] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1287.903225] env[61970]: value = "task-1356320" [ 1287.903225] env[61970]: _type = "Task" [ 1287.903225] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.903410] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6ae11223-2827-4b74-a2d5-b185efd513a3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.911692] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356320, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.912781] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1287.912781] env[61970]: value = "task-1356321" [ 1287.912781] env[61970]: _type = "Task" [ 1287.912781] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.919656] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356321, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.416609] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356320, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.424089] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356321, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.055162} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.424316] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1288.425062] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6f12bd-8b2f-4edf-96b9-fb74df62d85f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.449210] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 9d77a294-09ca-447c-924b-4feb0124dc70/9d77a294-09ca-447c-924b-4feb0124dc70.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1288.449487] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-feebdcdd-caeb-47fb-8e5f-e954417feba4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.467390] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1288.467390] env[61970]: value = "task-1356322" [ 1288.467390] env[61970]: _type = "Task" [ 1288.467390] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.474493] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356322, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.918230] env[61970]: DEBUG oslo_vmware.api [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356320, 'name': PowerOnVM_Task, 'duration_secs': 0.536404} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.918523] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1288.918721] env[61970]: INFO nova.compute.manager [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Took 7.67 seconds to spawn the instance on the hypervisor. [ 1288.918898] env[61970]: DEBUG nova.compute.manager [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1288.919662] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b84b5db-d20f-489b-a779-b51dfcb65d72 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.976636] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356322, 'name': ReconfigVM_Task, 'duration_secs': 0.267583} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.976892] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 9d77a294-09ca-447c-924b-4feb0124dc70/9d77a294-09ca-447c-924b-4feb0124dc70.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1288.977518] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f15240d0-631b-484c-9295-25cad4a17006 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.983297] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1288.983297] env[61970]: value = "task-1356323" [ 1288.983297] env[61970]: _type = "Task" [ 1288.983297] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.990326] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356323, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.438573] env[61970]: INFO nova.compute.manager [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Took 12.34 seconds to build instance. [ 1289.493259] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356323, 'name': Rename_Task, 'duration_secs': 0.129541} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.493510] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1289.493752] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e8f88ebb-cf67-4f00-8c10-7a389b1d2163 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.500131] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1289.500131] env[61970]: value = "task-1356324" [ 1289.500131] env[61970]: _type = "Task" [ 1289.500131] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.507202] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356324, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.941038] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ccd69c9a-3c15-4da3-83ad-cb7f6a5335ce tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "d468d634-e0ae-4407-beeb-4c56049ba203" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.852s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1290.010064] env[61970]: DEBUG oslo_vmware.api [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356324, 'name': PowerOnVM_Task, 'duration_secs': 0.429069} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1290.010391] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1290.010602] env[61970]: INFO nova.compute.manager [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Took 6.64 seconds to spawn the instance on the hypervisor. [ 1290.010789] env[61970]: DEBUG nova.compute.manager [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1290.011613] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829f7de9-c723-4a56-9d42-3cfe6fc8208b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.526954] env[61970]: INFO nova.compute.manager [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Took 13.12 seconds to build instance. [ 1291.028709] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5d073556-9d77-4f6b-95b8-b66cf010899d tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "9d77a294-09ca-447c-924b-4feb0124dc70" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.632s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1291.698677] env[61970]: INFO nova.compute.manager [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Rescuing [ 1291.698965] env[61970]: DEBUG oslo_concurrency.lockutils [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "refresh_cache-9d77a294-09ca-447c-924b-4feb0124dc70" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1291.699156] env[61970]: DEBUG oslo_concurrency.lockutils [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired lock "refresh_cache-9d77a294-09ca-447c-924b-4feb0124dc70" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1291.699333] env[61970]: DEBUG nova.network.neutron [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1292.430163] env[61970]: DEBUG nova.network.neutron [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Updating instance_info_cache with network_info: [{"id": "1e976667-f251-46e2-ae84-6b74c304cf63", "address": "fa:16:3e:cb:7d:1b", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e976667-f2", "ovs_interfaceid": "1e976667-f251-46e2-ae84-6b74c304cf63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1292.933741] env[61970]: DEBUG oslo_concurrency.lockutils [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Releasing lock "refresh_cache-9d77a294-09ca-447c-924b-4feb0124dc70" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1293.458810] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1293.459136] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c183a16e-a0be-4b3b-a784-7e9d8db74ce0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.466046] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1293.466046] env[61970]: value = "task-1356325" [ 1293.466046] env[61970]: _type = "Task" [ 1293.466046] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.474438] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356325, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.975749] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356325, 'name': PowerOffVM_Task, 'duration_secs': 0.199707} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.976038] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1293.976845] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a0b5da-5d07-46bb-931c-16b4bf82f431 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.995975] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b83f04ef-d0db-44e9-8926-03dee694ff60 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.027060] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1294.027379] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb4a46a0-6e66-486c-9924-001f4f225b90 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.034800] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1294.034800] env[61970]: value = "task-1356326" [ 1294.034800] env[61970]: _type = "Task" [ 1294.034800] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1294.042755] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356326, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.546330] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] VM already powered off {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1294.546715] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1294.546890] env[61970]: DEBUG oslo_concurrency.lockutils [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1294.547075] env[61970]: DEBUG oslo_concurrency.lockutils [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1294.547289] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1294.547585] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d3ac6c58-f917-4d7a-9056-4757ca9b71c5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.559649] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1294.559841] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1294.560739] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b452409-18ef-489f-a345-cb393acfcb51 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.565873] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1294.565873] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]524028e7-0cb4-dddc-f722-c267ec172bfe" [ 1294.565873] env[61970]: _type = "Task" [ 1294.565873] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1294.573640] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]524028e7-0cb4-dddc-f722-c267ec172bfe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.075988] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]524028e7-0cb4-dddc-f722-c267ec172bfe, 'name': SearchDatastore_Task, 'duration_secs': 0.009423} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1295.076710] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c80ad5b2-bbf6-4dff-a88e-930960436f14 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.081472] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1295.081472] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e02d10-43b4-ace5-9021-a24256e5e0b0" [ 1295.081472] env[61970]: _type = "Task" [ 1295.081472] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1295.088618] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e02d10-43b4-ace5-9021-a24256e5e0b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.593528] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52e02d10-43b4-ace5-9021-a24256e5e0b0, 'name': SearchDatastore_Task, 'duration_secs': 0.008676} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1295.593921] env[61970]: DEBUG oslo_concurrency.lockutils [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1295.594023] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 9d77a294-09ca-447c-924b-4feb0124dc70/7ff5af44-3f73-4c84-82e5-a1543db26d03-rescue.vmdk. {{(pid=61970) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1295.594288] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6708b200-2252-4761-82d6-348124ae3e2d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.601209] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1295.601209] env[61970]: value = "task-1356327" [ 1295.601209] env[61970]: _type = "Task" [ 1295.601209] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1295.608604] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356327, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.111950] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356327, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490682} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1296.112283] env[61970]: INFO nova.virt.vmwareapi.ds_util [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 9d77a294-09ca-447c-924b-4feb0124dc70/7ff5af44-3f73-4c84-82e5-a1543db26d03-rescue.vmdk. [ 1296.113140] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-038af636-d4ae-4c9f-980a-488b07602d4d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.136941] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 9d77a294-09ca-447c-924b-4feb0124dc70/7ff5af44-3f73-4c84-82e5-a1543db26d03-rescue.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1296.137232] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d247bb19-9060-4e41-a8f9-a07aece88e47 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.154767] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1296.154767] env[61970]: value = "task-1356328" [ 1296.154767] env[61970]: _type = "Task" [ 1296.154767] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1296.161803] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356328, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.664474] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356328, 'name': ReconfigVM_Task, 'duration_secs': 0.269602} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1296.664858] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 9d77a294-09ca-447c-924b-4feb0124dc70/7ff5af44-3f73-4c84-82e5-a1543db26d03-rescue.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1296.665567] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8aa0c35-141d-4d20-b7d0-566ace1be5d2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.689122] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-349d8c3d-faac-464b-923e-17987e77758a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.703012] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1296.703012] env[61970]: value = "task-1356329" [ 1296.703012] env[61970]: _type = "Task" [ 1296.703012] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1296.713278] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356329, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1297.212159] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356329, 'name': ReconfigVM_Task, 'duration_secs': 0.145422} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1297.212443] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1297.212703] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e24ecde1-c294-443a-b1fd-2a5a4e5b5a81 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.218562] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1297.218562] env[61970]: value = "task-1356330" [ 1297.218562] env[61970]: _type = "Task" [ 1297.218562] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1297.225240] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356330, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1297.729122] env[61970]: DEBUG oslo_vmware.api [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356330, 'name': PowerOnVM_Task, 'duration_secs': 0.403403} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1297.729547] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1297.732261] env[61970]: DEBUG nova.compute.manager [None req-aa09cb73-789f-4154-8134-94648bb03ec8 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1297.733045] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d80335cb-e1c5-4a83-84cd-5db9c5945681 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.823905] env[61970]: INFO nova.compute.manager [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Rescuing [ 1299.824397] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "refresh_cache-d468d634-e0ae-4407-beeb-4c56049ba203" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1299.824397] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired lock "refresh_cache-d468d634-e0ae-4407-beeb-4c56049ba203" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1299.824517] env[61970]: DEBUG nova.network.neutron [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1300.553872] env[61970]: DEBUG nova.network.neutron [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Updating instance_info_cache with network_info: [{"id": "1a5feac9-a368-4804-bb4a-648e682d6682", "address": "fa:16:3e:62:94:3e", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a5feac9-a3", "ovs_interfaceid": "1a5feac9-a368-4804-bb4a-648e682d6682", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1301.057023] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Releasing lock "refresh_cache-d468d634-e0ae-4407-beeb-4c56049ba203" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1301.586670] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1301.586983] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e8a9d929-0243-4f58-8d0f-b48f660c0fe0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.594157] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1301.594157] env[61970]: value = "task-1356331" [ 1301.594157] env[61970]: _type = "Task" [ 1301.594157] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1301.602415] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356331, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1302.104088] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356331, 'name': PowerOffVM_Task, 'duration_secs': 0.191856} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1302.104449] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1302.105161] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605d58ad-d80e-4828-a8d9-0ac7ce9a3c25 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.122856] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab9037bf-d5ad-45d6-b3ec-35c2d3ab602a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.148698] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1302.148961] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-df5098f1-393b-4cae-ba8d-2e39baccc167 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.155188] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1302.155188] env[61970]: value = "task-1356332" [ 1302.155188] env[61970]: _type = "Task" [ 1302.155188] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1302.162597] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356332, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1302.666462] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] VM already powered off {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1302.666688] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1302.666923] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1302.667088] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1302.667272] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1302.667534] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c27ec5d5-9413-482f-b204-f152bde6f9f1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.676153] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1302.676348] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1302.677074] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1aed4bd-1181-4529-9f50-0bbb430f9632 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.681953] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1302.681953] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5233ac31-8610-9fd7-2ce3-391f6caa5a0c" [ 1302.681953] env[61970]: _type = "Task" [ 1302.681953] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1302.690464] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5233ac31-8610-9fd7-2ce3-391f6caa5a0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.192182] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5233ac31-8610-9fd7-2ce3-391f6caa5a0c, 'name': SearchDatastore_Task, 'duration_secs': 0.007839} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.192934] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e467191-0b24-4cc9-8adc-2285320f266b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.197755] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1303.197755] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523478fa-195f-436b-67e7-2d8917952235" [ 1303.197755] env[61970]: _type = "Task" [ 1303.197755] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.204906] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523478fa-195f-436b-67e7-2d8917952235, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.708258] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523478fa-195f-436b-67e7-2d8917952235, 'name': SearchDatastore_Task, 'duration_secs': 0.009084} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.708548] env[61970]: DEBUG oslo_concurrency.lockutils [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1303.708849] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] d468d634-e0ae-4407-beeb-4c56049ba203/7ff5af44-3f73-4c84-82e5-a1543db26d03-rescue.vmdk. {{(pid=61970) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1303.709133] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-847b2f66-7b7e-42fa-8e77-13acfb92972b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.715286] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1303.715286] env[61970]: value = "task-1356333" [ 1303.715286] env[61970]: _type = "Task" [ 1303.715286] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.722536] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356333, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.225190] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356333, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.431125} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.225517] env[61970]: INFO nova.virt.vmwareapi.ds_util [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] d468d634-e0ae-4407-beeb-4c56049ba203/7ff5af44-3f73-4c84-82e5-a1543db26d03-rescue.vmdk. [ 1304.226205] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33fcc13e-2ea2-47dc-8990-2d59467a0417 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.249635] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] d468d634-e0ae-4407-beeb-4c56049ba203/7ff5af44-3f73-4c84-82e5-a1543db26d03-rescue.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1304.249866] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06483ee6-7a62-4771-b9ab-d28818d6bfe5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.266808] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1304.266808] env[61970]: value = "task-1356334" [ 1304.266808] env[61970]: _type = "Task" [ 1304.266808] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.273966] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356334, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.776511] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356334, 'name': ReconfigVM_Task, 'duration_secs': 0.295203} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.776746] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Reconfigured VM instance instance-0000006c to attach disk [datastore2] d468d634-e0ae-4407-beeb-4c56049ba203/7ff5af44-3f73-4c84-82e5-a1543db26d03-rescue.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1304.777597] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f514336-c5c2-4e8b-b263-f689f7454740 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.801634] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0cca3434-82e8-4a3f-92a9-af02c591221d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.816713] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1304.816713] env[61970]: value = "task-1356335" [ 1304.816713] env[61970]: _type = "Task" [ 1304.816713] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.824351] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356335, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.326453] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356335, 'name': ReconfigVM_Task, 'duration_secs': 0.250154} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1305.326851] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1305.326965] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7aeb54ff-ee1f-45d0-9f45-b0669fd99ef3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.333527] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1305.333527] env[61970]: value = "task-1356336" [ 1305.333527] env[61970]: _type = "Task" [ 1305.333527] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1305.341799] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356336, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.843408] env[61970]: DEBUG oslo_vmware.api [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356336, 'name': PowerOnVM_Task, 'duration_secs': 0.426248} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1305.843694] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1305.846506] env[61970]: DEBUG nova.compute.manager [None req-c413f885-9df0-4527-aa62-6d7cd3636cf2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1305.847271] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52540c35-85db-4cea-8ffe-88def8bddf24 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.672462] env[61970]: INFO nova.compute.manager [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Unrescuing [ 1306.672873] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "refresh_cache-d468d634-e0ae-4407-beeb-4c56049ba203" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1306.672917] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired lock "refresh_cache-d468d634-e0ae-4407-beeb-4c56049ba203" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1306.673103] env[61970]: DEBUG nova.network.neutron [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1307.361879] env[61970]: DEBUG nova.network.neutron [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Updating instance_info_cache with network_info: [{"id": "1a5feac9-a368-4804-bb4a-648e682d6682", "address": "fa:16:3e:62:94:3e", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a5feac9-a3", "ovs_interfaceid": "1a5feac9-a368-4804-bb4a-648e682d6682", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1307.864842] env[61970]: DEBUG oslo_concurrency.lockutils [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Releasing lock "refresh_cache-d468d634-e0ae-4407-beeb-4c56049ba203" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1307.865504] env[61970]: DEBUG nova.objects.instance [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lazy-loading 'flavor' on Instance uuid d468d634-e0ae-4407-beeb-4c56049ba203 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1308.372207] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f664d87b-3cf7-4c91-903c-da8939aae9b1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.392839] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1308.393166] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe529b54-0bf0-4460-9649-aeb442bf2b81 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.399189] env[61970]: DEBUG oslo_vmware.api [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1308.399189] env[61970]: value = "task-1356337" [ 1308.399189] env[61970]: _type = "Task" [ 1308.399189] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1308.406797] env[61970]: DEBUG oslo_vmware.api [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356337, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.909262] env[61970]: DEBUG oslo_vmware.api [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356337, 'name': PowerOffVM_Task, 'duration_secs': 0.183976} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.909699] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1308.914716] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Reconfiguring VM instance instance-0000006c to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1308.914987] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10dcb93b-0cbd-44ed-9932-192de61ac3dd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.933683] env[61970]: DEBUG oslo_vmware.api [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1308.933683] env[61970]: value = "task-1356338" [ 1308.933683] env[61970]: _type = "Task" [ 1308.933683] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1308.940744] env[61970]: DEBUG oslo_vmware.api [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356338, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1309.447689] env[61970]: DEBUG oslo_vmware.api [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356338, 'name': ReconfigVM_Task, 'duration_secs': 0.198217} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1309.448156] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Reconfigured VM instance instance-0000006c to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1309.448482] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1309.448979] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b64e271c-fdb2-4ee7-84e9-b6bf822b9531 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.455306] env[61970]: DEBUG oslo_vmware.api [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1309.455306] env[61970]: value = "task-1356339" [ 1309.455306] env[61970]: _type = "Task" [ 1309.455306] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1309.462484] env[61970]: DEBUG oslo_vmware.api [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356339, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1309.965470] env[61970]: DEBUG oslo_vmware.api [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356339, 'name': PowerOnVM_Task, 'duration_secs': 0.335375} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1309.965997] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1309.965997] env[61970]: DEBUG nova.compute.manager [None req-b2d1ec8b-cb98-44ec-b0bd-b80338d199e3 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1309.966896] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99bfdfc-d328-4e7a-8ba1-05b06ca4418d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.432484] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "2b6496dd-2efe-4767-908f-ff695435ef14" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1313.432790] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "2b6496dd-2efe-4767-908f-ff695435ef14" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1313.934913] env[61970]: DEBUG nova.compute.manager [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1314.458360] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1314.458701] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1314.460225] env[61970]: INFO nova.compute.claims [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1315.514042] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a46932ac-7c9e-4d0b-8e94-eec2c7ccc60f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.521019] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ee46b6d-a21f-4e7d-a79f-325f7f9d8ef6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.550429] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-840589d3-b9ec-4e1e-b203-384e0de7b6dd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.557230] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51203124-65f8-4905-a514-e6c5ac86f479 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.569660] env[61970]: DEBUG nova.compute.provider_tree [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1316.073252] env[61970]: DEBUG nova.scheduler.client.report [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1316.579060] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.120s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1316.579524] env[61970]: DEBUG nova.compute.manager [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1317.085052] env[61970]: DEBUG nova.compute.utils [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1317.086403] env[61970]: DEBUG nova.compute.manager [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1317.086559] env[61970]: DEBUG nova.network.neutron [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1317.139024] env[61970]: DEBUG nova.policy [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '248175c9d03f483eadaadab6bec2a3c9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b739147231ef473288ef6c78eed1caf3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 1317.383460] env[61970]: DEBUG nova.network.neutron [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Successfully created port: 7b4d4245-6fe4-4673-aede-924dc3cdce97 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1317.590237] env[61970]: DEBUG nova.compute.manager [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1318.601083] env[61970]: DEBUG nova.compute.manager [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1318.628021] env[61970]: DEBUG nova.virt.hardware [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1318.628021] env[61970]: DEBUG nova.virt.hardware [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1318.628021] env[61970]: DEBUG nova.virt.hardware [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1318.628021] env[61970]: DEBUG nova.virt.hardware [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1318.628290] env[61970]: DEBUG nova.virt.hardware [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1318.628343] env[61970]: DEBUG nova.virt.hardware [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1318.628551] env[61970]: DEBUG nova.virt.hardware [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1318.628719] env[61970]: DEBUG nova.virt.hardware [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1318.628896] env[61970]: DEBUG nova.virt.hardware [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1318.629409] env[61970]: DEBUG nova.virt.hardware [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1318.629409] env[61970]: DEBUG nova.virt.hardware [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1318.630147] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f0cf7b-267f-4f4f-ab7d-44fd40270ecb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.638289] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ab0daf-8913-48ab-b42c-0c2fcefbac5b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.745243] env[61970]: DEBUG nova.compute.manager [req-7c5c02c6-720c-43ca-8ec0-f837850a2164 req-e5745e12-edd9-4915-b990-f78522bd2a37 service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Received event network-vif-plugged-7b4d4245-6fe4-4673-aede-924dc3cdce97 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1318.745518] env[61970]: DEBUG oslo_concurrency.lockutils [req-7c5c02c6-720c-43ca-8ec0-f837850a2164 req-e5745e12-edd9-4915-b990-f78522bd2a37 service nova] Acquiring lock "2b6496dd-2efe-4767-908f-ff695435ef14-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1318.745694] env[61970]: DEBUG oslo_concurrency.lockutils [req-7c5c02c6-720c-43ca-8ec0-f837850a2164 req-e5745e12-edd9-4915-b990-f78522bd2a37 service nova] Lock "2b6496dd-2efe-4767-908f-ff695435ef14-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1318.745883] env[61970]: DEBUG oslo_concurrency.lockutils [req-7c5c02c6-720c-43ca-8ec0-f837850a2164 req-e5745e12-edd9-4915-b990-f78522bd2a37 service nova] Lock "2b6496dd-2efe-4767-908f-ff695435ef14-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1318.746015] env[61970]: DEBUG nova.compute.manager [req-7c5c02c6-720c-43ca-8ec0-f837850a2164 req-e5745e12-edd9-4915-b990-f78522bd2a37 service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] No waiting events found dispatching network-vif-plugged-7b4d4245-6fe4-4673-aede-924dc3cdce97 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1318.746837] env[61970]: WARNING nova.compute.manager [req-7c5c02c6-720c-43ca-8ec0-f837850a2164 req-e5745e12-edd9-4915-b990-f78522bd2a37 service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Received unexpected event network-vif-plugged-7b4d4245-6fe4-4673-aede-924dc3cdce97 for instance with vm_state building and task_state spawning. [ 1318.831352] env[61970]: DEBUG nova.network.neutron [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Successfully updated port: 7b4d4245-6fe4-4673-aede-924dc3cdce97 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1319.333559] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1319.333733] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1319.333887] env[61970]: DEBUG nova.network.neutron [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1319.864879] env[61970]: DEBUG nova.network.neutron [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1319.979614] env[61970]: DEBUG nova.network.neutron [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Updating instance_info_cache with network_info: [{"id": "7b4d4245-6fe4-4673-aede-924dc3cdce97", "address": "fa:16:3e:81:9c:2a", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b4d4245-6f", "ovs_interfaceid": "7b4d4245-6fe4-4673-aede-924dc3cdce97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1320.482600] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Releasing lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1320.482966] env[61970]: DEBUG nova.compute.manager [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Instance network_info: |[{"id": "7b4d4245-6fe4-4673-aede-924dc3cdce97", "address": "fa:16:3e:81:9c:2a", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b4d4245-6f", "ovs_interfaceid": "7b4d4245-6fe4-4673-aede-924dc3cdce97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1320.483700] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:9c:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '669e4919-e0ad-4e23-9f23-4c5f2be0d858', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7b4d4245-6fe4-4673-aede-924dc3cdce97', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1320.491220] env[61970]: DEBUG oslo.service.loopingcall [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1320.491564] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1320.492188] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-00543bda-88e5-4417-9b4e-33122e2c4738 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.511976] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1320.511976] env[61970]: value = "task-1356340" [ 1320.511976] env[61970]: _type = "Task" [ 1320.511976] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.519385] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356340, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.770270] env[61970]: DEBUG nova.compute.manager [req-abc1de56-6134-4e4b-a395-486f1709a460 req-c8acafac-2979-40ce-9826-193a93b8111d service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Received event network-changed-7b4d4245-6fe4-4673-aede-924dc3cdce97 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1320.770523] env[61970]: DEBUG nova.compute.manager [req-abc1de56-6134-4e4b-a395-486f1709a460 req-c8acafac-2979-40ce-9826-193a93b8111d service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Refreshing instance network info cache due to event network-changed-7b4d4245-6fe4-4673-aede-924dc3cdce97. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1320.770790] env[61970]: DEBUG oslo_concurrency.lockutils [req-abc1de56-6134-4e4b-a395-486f1709a460 req-c8acafac-2979-40ce-9826-193a93b8111d service nova] Acquiring lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1320.770977] env[61970]: DEBUG oslo_concurrency.lockutils [req-abc1de56-6134-4e4b-a395-486f1709a460 req-c8acafac-2979-40ce-9826-193a93b8111d service nova] Acquired lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1320.771356] env[61970]: DEBUG nova.network.neutron [req-abc1de56-6134-4e4b-a395-486f1709a460 req-c8acafac-2979-40ce-9826-193a93b8111d service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Refreshing network info cache for port 7b4d4245-6fe4-4673-aede-924dc3cdce97 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1321.021859] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356340, 'name': CreateVM_Task, 'duration_secs': 0.308065} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1321.022211] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1321.022663] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1321.022838] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1321.023176] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1321.023439] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ec29bd2-96ca-4316-8608-b29993956e04 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.027695] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1321.027695] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ac2199-c15e-28c7-ec9b-ab7bc48506e3" [ 1321.027695] env[61970]: _type = "Task" [ 1321.027695] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1321.034977] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ac2199-c15e-28c7-ec9b-ab7bc48506e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1321.446800] env[61970]: DEBUG nova.network.neutron [req-abc1de56-6134-4e4b-a395-486f1709a460 req-c8acafac-2979-40ce-9826-193a93b8111d service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Updated VIF entry in instance network info cache for port 7b4d4245-6fe4-4673-aede-924dc3cdce97. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1321.447184] env[61970]: DEBUG nova.network.neutron [req-abc1de56-6134-4e4b-a395-486f1709a460 req-c8acafac-2979-40ce-9826-193a93b8111d service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Updating instance_info_cache with network_info: [{"id": "7b4d4245-6fe4-4673-aede-924dc3cdce97", "address": "fa:16:3e:81:9c:2a", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b4d4245-6f", "ovs_interfaceid": "7b4d4245-6fe4-4673-aede-924dc3cdce97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1321.539526] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ac2199-c15e-28c7-ec9b-ab7bc48506e3, 'name': SearchDatastore_Task, 'duration_secs': 0.011047} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1321.539826] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1321.540074] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1321.540312] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1321.540487] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1321.540671] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1321.540923] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6d36f454-ac0e-489f-9509-c35e90e570bf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.548338] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1321.548515] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1321.549185] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc1800db-56d4-4d19-b6c0-ad3622b8d282 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.553773] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1321.553773] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52704b95-c8e5-6f18-88d5-ea093eb04610" [ 1321.553773] env[61970]: _type = "Task" [ 1321.553773] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1321.560655] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52704b95-c8e5-6f18-88d5-ea093eb04610, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1321.950526] env[61970]: DEBUG oslo_concurrency.lockutils [req-abc1de56-6134-4e4b-a395-486f1709a460 req-c8acafac-2979-40ce-9826-193a93b8111d service nova] Releasing lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1322.064200] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52704b95-c8e5-6f18-88d5-ea093eb04610, 'name': SearchDatastore_Task, 'duration_secs': 0.007148} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1322.064962] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad833864-3637-4347-b194-4250365c3d0d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.069882] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1322.069882] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]527eadea-9d2c-8aa8-9ac2-19881f5d7fd1" [ 1322.069882] env[61970]: _type = "Task" [ 1322.069882] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.077592] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]527eadea-9d2c-8aa8-9ac2-19881f5d7fd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1322.580602] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]527eadea-9d2c-8aa8-9ac2-19881f5d7fd1, 'name': SearchDatastore_Task, 'duration_secs': 0.009163} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1322.580862] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1322.581141] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] 2b6496dd-2efe-4767-908f-ff695435ef14/2b6496dd-2efe-4767-908f-ff695435ef14.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1322.581418] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-45317d31-147e-44a2-ae36-0fb939591ff0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.588212] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1322.588212] env[61970]: value = "task-1356341" [ 1322.588212] env[61970]: _type = "Task" [ 1322.588212] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.595496] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356341, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.098805] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356341, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.378111} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1323.099231] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] 2b6496dd-2efe-4767-908f-ff695435ef14/2b6496dd-2efe-4767-908f-ff695435ef14.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1323.099551] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1323.099824] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d9d00119-43b7-43b2-99d9-06560ec02d45 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.106154] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1323.106154] env[61970]: value = "task-1356342" [ 1323.106154] env[61970]: _type = "Task" [ 1323.106154] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1323.113646] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356342, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.615681] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356342, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07574} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1323.615951] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1323.616822] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f932fe88-d251-4a6c-ba71-cb27a05494e4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.638059] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 2b6496dd-2efe-4767-908f-ff695435ef14/2b6496dd-2efe-4767-908f-ff695435ef14.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1323.638059] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90576023-b88b-4504-9b04-764dabf74a36 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.656849] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1323.656849] env[61970]: value = "task-1356343" [ 1323.656849] env[61970]: _type = "Task" [ 1323.656849] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1323.664238] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356343, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.167090] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356343, 'name': ReconfigVM_Task, 'duration_secs': 0.26282} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1324.167395] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 2b6496dd-2efe-4767-908f-ff695435ef14/2b6496dd-2efe-4767-908f-ff695435ef14.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1324.168055] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-72873ba4-a0ea-4b1f-8dbe-565700ea4069 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.173973] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1324.173973] env[61970]: value = "task-1356344" [ 1324.173973] env[61970]: _type = "Task" [ 1324.173973] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1324.183319] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356344, 'name': Rename_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.684365] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356344, 'name': Rename_Task} progress is 14%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.185807] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356344, 'name': Rename_Task, 'duration_secs': 0.5882} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1325.187260] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1325.187260] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bb9536c8-9bea-47f9-8349-2d0629305573 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.192280] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1325.192280] env[61970]: value = "task-1356345" [ 1325.192280] env[61970]: _type = "Task" [ 1325.192280] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1325.199043] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356345, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.259771] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1325.701695] env[61970]: DEBUG oslo_vmware.api [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356345, 'name': PowerOnVM_Task, 'duration_secs': 0.42526} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1325.701995] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1325.702225] env[61970]: INFO nova.compute.manager [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Took 7.10 seconds to spawn the instance on the hypervisor. [ 1325.702407] env[61970]: DEBUG nova.compute.manager [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1325.703157] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4b01a2c-ae30-4b4b-8441-54c5f5d48ae5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.222205] env[61970]: INFO nova.compute.manager [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Took 11.78 seconds to build instance. [ 1326.583454] env[61970]: DEBUG nova.compute.manager [req-efa98575-141f-4b33-b49c-ad01fb51bfb7 req-e97f1dc4-7084-43e1-9764-23c481c5d5fe service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Received event network-changed-7b4d4245-6fe4-4673-aede-924dc3cdce97 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1326.583562] env[61970]: DEBUG nova.compute.manager [req-efa98575-141f-4b33-b49c-ad01fb51bfb7 req-e97f1dc4-7084-43e1-9764-23c481c5d5fe service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Refreshing instance network info cache due to event network-changed-7b4d4245-6fe4-4673-aede-924dc3cdce97. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1326.584546] env[61970]: DEBUG oslo_concurrency.lockutils [req-efa98575-141f-4b33-b49c-ad01fb51bfb7 req-e97f1dc4-7084-43e1-9764-23c481c5d5fe service nova] Acquiring lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1326.584546] env[61970]: DEBUG oslo_concurrency.lockutils [req-efa98575-141f-4b33-b49c-ad01fb51bfb7 req-e97f1dc4-7084-43e1-9764-23c481c5d5fe service nova] Acquired lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1326.584546] env[61970]: DEBUG nova.network.neutron [req-efa98575-141f-4b33-b49c-ad01fb51bfb7 req-e97f1dc4-7084-43e1-9764-23c481c5d5fe service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Refreshing network info cache for port 7b4d4245-6fe4-4673-aede-924dc3cdce97 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1326.727880] env[61970]: DEBUG oslo_concurrency.lockutils [None req-f4ae213e-2f91-4d0f-9f49-5f379ff5b00a tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "2b6496dd-2efe-4767-908f-ff695435ef14" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.295s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1327.260837] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1327.296170] env[61970]: DEBUG nova.network.neutron [req-efa98575-141f-4b33-b49c-ad01fb51bfb7 req-e97f1dc4-7084-43e1-9764-23c481c5d5fe service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Updated VIF entry in instance network info cache for port 7b4d4245-6fe4-4673-aede-924dc3cdce97. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1327.296552] env[61970]: DEBUG nova.network.neutron [req-efa98575-141f-4b33-b49c-ad01fb51bfb7 req-e97f1dc4-7084-43e1-9764-23c481c5d5fe service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Updating instance_info_cache with network_info: [{"id": "7b4d4245-6fe4-4673-aede-924dc3cdce97", "address": "fa:16:3e:81:9c:2a", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.131", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b4d4245-6f", "ovs_interfaceid": "7b4d4245-6fe4-4673-aede-924dc3cdce97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1327.799113] env[61970]: DEBUG oslo_concurrency.lockutils [req-efa98575-141f-4b33-b49c-ad01fb51bfb7 req-e97f1dc4-7084-43e1-9764-23c481c5d5fe service nova] Releasing lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1331.255946] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1331.259604] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1331.763051] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1331.763309] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1331.763487] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1331.763647] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61970) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1331.764601] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35970fb8-279d-461a-bd82-225dd4c8b167 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.772921] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-586c168d-63a2-4a84-a259-c76a94f3f7f4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.786767] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-359575e7-071a-4811-9ce9-ec515c749f7a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.792927] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f879df-f526-4789-a7f1-dfa17e3240f9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.821785] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181493MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61970) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1331.821925] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1331.822117] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1332.849026] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance d468d634-e0ae-4407-beeb-4c56049ba203 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1332.849026] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 9d77a294-09ca-447c-924b-4feb0124dc70 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1332.849026] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 2b6496dd-2efe-4767-908f-ff695435ef14 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1332.849026] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1332.849026] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1332.900670] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77a6e481-7af6-4618-a781-b91c6b5144dd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.908080] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc86e4cc-c27d-42e7-803c-8464829c7936 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.936880] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b14a2b-29f0-4ed7-ac31-4e028dc4ebb4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.943625] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df29a96-89de-43ee-97c3-90ff8a09a910 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.957199] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1333.460330] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1333.965657] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1333.966103] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.144s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1334.966026] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1335.471398] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1335.471575] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Starting heal instance info cache {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1335.471677] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Rebuilding the list of instances to heal {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1336.006027] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "refresh_cache-d468d634-e0ae-4407-beeb-4c56049ba203" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1336.006288] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquired lock "refresh_cache-d468d634-e0ae-4407-beeb-4c56049ba203" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1336.006332] env[61970]: DEBUG nova.network.neutron [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Forcefully refreshing network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1336.006480] env[61970]: DEBUG nova.objects.instance [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lazy-loading 'info_cache' on Instance uuid d468d634-e0ae-4407-beeb-4c56049ba203 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1337.696380] env[61970]: DEBUG nova.network.neutron [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Updating instance_info_cache with network_info: [{"id": "1a5feac9-a368-4804-bb4a-648e682d6682", "address": "fa:16:3e:62:94:3e", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a5feac9-a3", "ovs_interfaceid": "1a5feac9-a368-4804-bb4a-648e682d6682", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1338.199146] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Releasing lock "refresh_cache-d468d634-e0ae-4407-beeb-4c56049ba203" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1338.199376] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Updated the network info_cache for instance {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1338.199579] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1338.199737] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1338.199891] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1338.200050] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1338.200181] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61970) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1362.996366] env[61970]: DEBUG oslo_concurrency.lockutils [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "2b6496dd-2efe-4767-908f-ff695435ef14" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1362.996650] env[61970]: DEBUG oslo_concurrency.lockutils [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "2b6496dd-2efe-4767-908f-ff695435ef14" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1363.500385] env[61970]: DEBUG nova.compute.utils [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1364.004458] env[61970]: DEBUG oslo_concurrency.lockutils [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "2b6496dd-2efe-4767-908f-ff695435ef14" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1365.074656] env[61970]: DEBUG oslo_concurrency.lockutils [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "2b6496dd-2efe-4767-908f-ff695435ef14" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1365.075312] env[61970]: DEBUG oslo_concurrency.lockutils [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "2b6496dd-2efe-4767-908f-ff695435ef14" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1365.075568] env[61970]: INFO nova.compute.manager [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Attaching volume 2c5f4217-7dd6-4daf-81b9-d5443ce7ec59 to /dev/sdb [ 1365.104290] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-990075cc-0e7f-4b40-8d40-2295a7edc99c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.111005] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f18eac0-07ef-4c89-b0b7-88c2301eae40 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.123566] env[61970]: DEBUG nova.virt.block_device [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Updating existing volume attachment record: a7be389f-88f0-4983-b2c6-00fc7ebc9355 {{(pid=61970) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1369.668022] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Volume attach. Driver type: vmdk {{(pid=61970) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1369.668415] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288905', 'volume_id': '2c5f4217-7dd6-4daf-81b9-d5443ce7ec59', 'name': 'volume-2c5f4217-7dd6-4daf-81b9-d5443ce7ec59', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2b6496dd-2efe-4767-908f-ff695435ef14', 'attached_at': '', 'detached_at': '', 'volume_id': '2c5f4217-7dd6-4daf-81b9-d5443ce7ec59', 'serial': '2c5f4217-7dd6-4daf-81b9-d5443ce7ec59'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1369.669188] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499ad06c-a228-49a8-8ba4-9a79d4a414ff {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1369.685087] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a562eae-e18d-4c68-82da-af1113bff3cd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1369.708280] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] volume-2c5f4217-7dd6-4daf-81b9-d5443ce7ec59/volume-2c5f4217-7dd6-4daf-81b9-d5443ce7ec59.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1369.708534] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5854a403-ec88-4e92-b4db-a12cc07e5119 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1369.726826] env[61970]: DEBUG oslo_vmware.api [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1369.726826] env[61970]: value = "task-1356350" [ 1369.726826] env[61970]: _type = "Task" [ 1369.726826] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1369.734971] env[61970]: DEBUG oslo_vmware.api [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356350, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1370.236442] env[61970]: DEBUG oslo_vmware.api [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356350, 'name': ReconfigVM_Task, 'duration_secs': 0.340424} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1370.236724] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Reconfigured VM instance instance-0000006e to attach disk [datastore2] volume-2c5f4217-7dd6-4daf-81b9-d5443ce7ec59/volume-2c5f4217-7dd6-4daf-81b9-d5443ce7ec59.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1370.241254] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a61b78f-2746-454c-8439-d23cecf00a31 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.254854] env[61970]: DEBUG oslo_vmware.api [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1370.254854] env[61970]: value = "task-1356351" [ 1370.254854] env[61970]: _type = "Task" [ 1370.254854] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1370.263244] env[61970]: DEBUG oslo_vmware.api [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356351, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1370.764790] env[61970]: DEBUG oslo_vmware.api [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356351, 'name': ReconfigVM_Task, 'duration_secs': 0.131236} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1370.765123] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288905', 'volume_id': '2c5f4217-7dd6-4daf-81b9-d5443ce7ec59', 'name': 'volume-2c5f4217-7dd6-4daf-81b9-d5443ce7ec59', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2b6496dd-2efe-4767-908f-ff695435ef14', 'attached_at': '', 'detached_at': '', 'volume_id': '2c5f4217-7dd6-4daf-81b9-d5443ce7ec59', 'serial': '2c5f4217-7dd6-4daf-81b9-d5443ce7ec59'} {{(pid=61970) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1371.799822] env[61970]: DEBUG nova.objects.instance [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lazy-loading 'flavor' on Instance uuid 2b6496dd-2efe-4767-908f-ff695435ef14 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1372.305651] env[61970]: DEBUG oslo_concurrency.lockutils [None req-13b473d3-e4c6-41bb-9826-8ed40916104b tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "2b6496dd-2efe-4767-908f-ff695435ef14" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.230s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1372.440455] env[61970]: INFO nova.compute.manager [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Rescuing [ 1372.440729] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1372.440886] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1372.441068] env[61970]: DEBUG nova.network.neutron [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1373.155260] env[61970]: DEBUG nova.network.neutron [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Updating instance_info_cache with network_info: [{"id": "7b4d4245-6fe4-4673-aede-924dc3cdce97", "address": "fa:16:3e:81:9c:2a", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.131", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b4d4245-6f", "ovs_interfaceid": "7b4d4245-6fe4-4673-aede-924dc3cdce97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1373.658600] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Releasing lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1374.190757] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1374.191163] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3448c18e-c8d2-4e88-abe8-f3d14d419e07 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.198879] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1374.198879] env[61970]: value = "task-1356352" [ 1374.198879] env[61970]: _type = "Task" [ 1374.198879] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1374.207534] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356352, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1374.709310] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356352, 'name': PowerOffVM_Task, 'duration_secs': 0.188102} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1374.709577] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1374.710333] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c21ae04-56b3-43c9-86c3-3b4b15bbedc7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.731197] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4036a80-86af-40e5-865f-9a2e59405dc1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.756288] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1374.756531] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-be7b94ea-c496-46c9-998a-c8b7047d0fd8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.762235] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1374.762235] env[61970]: value = "task-1356353" [ 1374.762235] env[61970]: _type = "Task" [ 1374.762235] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1374.769077] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356353, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1375.272454] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] VM already powered off {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1375.272851] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1375.273054] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1375.273218] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1375.273419] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1375.273659] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b8b1caae-84e9-4795-8260-a2a619690c37 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.282887] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1375.283074] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1375.283906] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f9b7ff2-8ece-463a-8d35-b29d00dfa0ba {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.288814] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1375.288814] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523f4c80-8195-3e3d-dea9-9d7ce62b9453" [ 1375.288814] env[61970]: _type = "Task" [ 1375.288814] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1375.296167] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523f4c80-8195-3e3d-dea9-9d7ce62b9453, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1375.798894] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]523f4c80-8195-3e3d-dea9-9d7ce62b9453, 'name': SearchDatastore_Task, 'duration_secs': 0.008775} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1375.799747] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1ba769d-a9b9-490e-bf80-dce7ddb39b0f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.804606] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1375.804606] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]522a3b8f-bcbd-557b-386a-965d351c0fa6" [ 1375.804606] env[61970]: _type = "Task" [ 1375.804606] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1375.811431] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]522a3b8f-bcbd-557b-386a-965d351c0fa6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1376.315271] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]522a3b8f-bcbd-557b-386a-965d351c0fa6, 'name': SearchDatastore_Task, 'duration_secs': 0.008871} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1376.315650] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1376.315814] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] 2b6496dd-2efe-4767-908f-ff695435ef14/7ff5af44-3f73-4c84-82e5-a1543db26d03-rescue.vmdk. {{(pid=61970) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1376.316082] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-76b6ffe2-c5b4-4423-860d-c775ea7f7b83 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1376.322371] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1376.322371] env[61970]: value = "task-1356354" [ 1376.322371] env[61970]: _type = "Task" [ 1376.322371] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1376.329519] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356354, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1376.832118] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356354, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.384547} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1376.832439] env[61970]: INFO nova.virt.vmwareapi.ds_util [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore1] 2b6496dd-2efe-4767-908f-ff695435ef14/7ff5af44-3f73-4c84-82e5-a1543db26d03-rescue.vmdk. [ 1376.833239] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1006927-6507-4013-a9f2-00cba557ca79 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1376.859398] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 2b6496dd-2efe-4767-908f-ff695435ef14/7ff5af44-3f73-4c84-82e5-a1543db26d03-rescue.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1376.859655] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46b135aa-ef40-45f2-a95f-a61603acc275 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1376.876809] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1376.876809] env[61970]: value = "task-1356355" [ 1376.876809] env[61970]: _type = "Task" [ 1376.876809] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1376.883931] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356355, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1377.386469] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356355, 'name': ReconfigVM_Task, 'duration_secs': 0.288541} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1377.386809] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 2b6496dd-2efe-4767-908f-ff695435ef14/7ff5af44-3f73-4c84-82e5-a1543db26d03-rescue.vmdk or device None with type thin {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1377.387573] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4884fdd3-12d2-445b-82f6-e1b4fb0df727 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.415193] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-449c409d-d6b6-43c1-94f3-14b5f08ab40f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.429898] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1377.429898] env[61970]: value = "task-1356356" [ 1377.429898] env[61970]: _type = "Task" [ 1377.429898] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1377.436903] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356356, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1377.939908] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356356, 'name': ReconfigVM_Task, 'duration_secs': 0.163557} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1377.940190] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1377.940436] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ceba0d1c-70bb-4690-ab81-037cc0e02a41 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.945934] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1377.945934] env[61970]: value = "task-1356357" [ 1377.945934] env[61970]: _type = "Task" [ 1377.945934] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1377.952777] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356357, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1378.455729] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356357, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1378.956922] env[61970]: DEBUG oslo_vmware.api [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356357, 'name': PowerOnVM_Task, 'duration_secs': 1.002774} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1378.957223] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1378.960298] env[61970]: DEBUG nova.compute.manager [None req-ebb48c79-177b-48f4-80b1-f3d93489499c tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1378.961086] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3fb611-2432-4e0e-b226-6b0c5959f00f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.480456] env[61970]: INFO nova.compute.manager [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Unrescuing [ 1380.480886] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1380.480886] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquired lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1380.481109] env[61970]: DEBUG nova.network.neutron [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1381.211978] env[61970]: DEBUG nova.network.neutron [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Updating instance_info_cache with network_info: [{"id": "7b4d4245-6fe4-4673-aede-924dc3cdce97", "address": "fa:16:3e:81:9c:2a", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.131", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b4d4245-6f", "ovs_interfaceid": "7b4d4245-6fe4-4673-aede-924dc3cdce97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1381.715260] env[61970]: DEBUG oslo_concurrency.lockutils [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Releasing lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1381.715924] env[61970]: DEBUG nova.objects.instance [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lazy-loading 'flavor' on Instance uuid 2b6496dd-2efe-4767-908f-ff695435ef14 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1382.222335] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50621b5c-c323-49f0-9df5-ecdb7bb2fef5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1382.247631] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1382.247969] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4282324b-e12a-43a1-aed7-29397d6674a8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1382.255313] env[61970]: DEBUG oslo_vmware.api [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1382.255313] env[61970]: value = "task-1356358" [ 1382.255313] env[61970]: _type = "Task" [ 1382.255313] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1382.262861] env[61970]: DEBUG oslo_vmware.api [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356358, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1382.764799] env[61970]: DEBUG oslo_vmware.api [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356358, 'name': PowerOffVM_Task, 'duration_secs': 0.193384} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1382.765209] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1382.770139] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Reconfiguring VM instance instance-0000006e to detach disk 2002 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1382.770396] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-841c9332-67a3-475b-917c-53c9aac8954b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1382.787776] env[61970]: DEBUG oslo_vmware.api [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1382.787776] env[61970]: value = "task-1356359" [ 1382.787776] env[61970]: _type = "Task" [ 1382.787776] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1382.795175] env[61970]: DEBUG oslo_vmware.api [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356359, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1383.298696] env[61970]: DEBUG oslo_vmware.api [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356359, 'name': ReconfigVM_Task, 'duration_secs': 0.215753} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1383.299015] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Reconfigured VM instance instance-0000006e to detach disk 2002 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1383.299213] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1383.299465] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-87b7cf90-95f1-4733-a13c-2664ff570386 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.305579] env[61970]: DEBUG oslo_vmware.api [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1383.305579] env[61970]: value = "task-1356360" [ 1383.305579] env[61970]: _type = "Task" [ 1383.305579] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1383.312985] env[61970]: DEBUG oslo_vmware.api [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356360, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1383.814502] env[61970]: DEBUG oslo_vmware.api [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356360, 'name': PowerOnVM_Task, 'duration_secs': 0.35067} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1383.814896] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1383.815025] env[61970]: DEBUG nova.compute.manager [None req-8979bdd1-31f2-4691-9e21-7c4745d388c9 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1383.815746] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05002792-ae66-4b76-a053-994298359c54 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.141145] env[61970]: DEBUG nova.compute.manager [req-ab1e8321-2ee4-4029-90cc-00801828c6c2 req-3bbc45f8-4bc6-442c-9969-e13feb2c6762 service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Received event network-changed-7b4d4245-6fe4-4673-aede-924dc3cdce97 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1385.141403] env[61970]: DEBUG nova.compute.manager [req-ab1e8321-2ee4-4029-90cc-00801828c6c2 req-3bbc45f8-4bc6-442c-9969-e13feb2c6762 service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Refreshing instance network info cache due to event network-changed-7b4d4245-6fe4-4673-aede-924dc3cdce97. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1385.141811] env[61970]: DEBUG oslo_concurrency.lockutils [req-ab1e8321-2ee4-4029-90cc-00801828c6c2 req-3bbc45f8-4bc6-442c-9969-e13feb2c6762 service nova] Acquiring lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1385.141888] env[61970]: DEBUG oslo_concurrency.lockutils [req-ab1e8321-2ee4-4029-90cc-00801828c6c2 req-3bbc45f8-4bc6-442c-9969-e13feb2c6762 service nova] Acquired lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1385.142042] env[61970]: DEBUG nova.network.neutron [req-ab1e8321-2ee4-4029-90cc-00801828c6c2 req-3bbc45f8-4bc6-442c-9969-e13feb2c6762 service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Refreshing network info cache for port 7b4d4245-6fe4-4673-aede-924dc3cdce97 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1385.260277] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1385.847259] env[61970]: DEBUG nova.network.neutron [req-ab1e8321-2ee4-4029-90cc-00801828c6c2 req-3bbc45f8-4bc6-442c-9969-e13feb2c6762 service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Updated VIF entry in instance network info cache for port 7b4d4245-6fe4-4673-aede-924dc3cdce97. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1385.847801] env[61970]: DEBUG nova.network.neutron [req-ab1e8321-2ee4-4029-90cc-00801828c6c2 req-3bbc45f8-4bc6-442c-9969-e13feb2c6762 service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Updating instance_info_cache with network_info: [{"id": "7b4d4245-6fe4-4673-aede-924dc3cdce97", "address": "fa:16:3e:81:9c:2a", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.131", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b4d4245-6f", "ovs_interfaceid": "7b4d4245-6fe4-4673-aede-924dc3cdce97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1386.351040] env[61970]: DEBUG oslo_concurrency.lockutils [req-ab1e8321-2ee4-4029-90cc-00801828c6c2 req-3bbc45f8-4bc6-442c-9969-e13feb2c6762 service nova] Releasing lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1387.165512] env[61970]: DEBUG nova.compute.manager [req-e8f1369e-08a6-4f94-a524-222a2b90f967 req-0497b055-bce2-4966-a7f9-e35ece584d09 service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Received event network-changed-7b4d4245-6fe4-4673-aede-924dc3cdce97 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1387.165711] env[61970]: DEBUG nova.compute.manager [req-e8f1369e-08a6-4f94-a524-222a2b90f967 req-0497b055-bce2-4966-a7f9-e35ece584d09 service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Refreshing instance network info cache due to event network-changed-7b4d4245-6fe4-4673-aede-924dc3cdce97. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1387.165925] env[61970]: DEBUG oslo_concurrency.lockutils [req-e8f1369e-08a6-4f94-a524-222a2b90f967 req-0497b055-bce2-4966-a7f9-e35ece584d09 service nova] Acquiring lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1387.166090] env[61970]: DEBUG oslo_concurrency.lockutils [req-e8f1369e-08a6-4f94-a524-222a2b90f967 req-0497b055-bce2-4966-a7f9-e35ece584d09 service nova] Acquired lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1387.166257] env[61970]: DEBUG nova.network.neutron [req-e8f1369e-08a6-4f94-a524-222a2b90f967 req-0497b055-bce2-4966-a7f9-e35ece584d09 service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Refreshing network info cache for port 7b4d4245-6fe4-4673-aede-924dc3cdce97 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1387.854837] env[61970]: DEBUG nova.network.neutron [req-e8f1369e-08a6-4f94-a524-222a2b90f967 req-0497b055-bce2-4966-a7f9-e35ece584d09 service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Updated VIF entry in instance network info cache for port 7b4d4245-6fe4-4673-aede-924dc3cdce97. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1387.855226] env[61970]: DEBUG nova.network.neutron [req-e8f1369e-08a6-4f94-a524-222a2b90f967 req-0497b055-bce2-4966-a7f9-e35ece584d09 service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Updating instance_info_cache with network_info: [{"id": "7b4d4245-6fe4-4673-aede-924dc3cdce97", "address": "fa:16:3e:81:9c:2a", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.131", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b4d4245-6f", "ovs_interfaceid": "7b4d4245-6fe4-4673-aede-924dc3cdce97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1388.259906] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1388.260093] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Cleaning up deleted instances {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1388.358120] env[61970]: DEBUG oslo_concurrency.lockutils [req-e8f1369e-08a6-4f94-a524-222a2b90f967 req-0497b055-bce2-4966-a7f9-e35ece584d09 service nova] Releasing lock "refresh_cache-2b6496dd-2efe-4767-908f-ff695435ef14" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1388.765134] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] There are 8 instances to clean {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1388.765332] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: bead8709-273f-433f-9d06-4c01510aaca1] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1389.268753] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: a8d4f3f0-ccad-47b2-8232-a3f68f77a1d8] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1389.771787] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: df0a2338-07e0-4148-8729-a290b88207b3] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1390.275646] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 5a8cdb36-1845-4806-a6bc-22afb47886e2] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1390.779846] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 183133c6-7e4e-4dd4-b8e0-cb4f8338264c] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1391.283136] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 3c9d8e79-973a-4dc4-a59c-a3c96f5ea9c6] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1391.786128] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: d96208cf-fcc5-4014-ac23-1d56ea97050a] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1392.289832] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 0304640a-6c46-4f57-a8b8-5058da2e3de4] Instance has had 0 of 5 cleanup attempts {{(pid=61970) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1393.793616] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1393.794076] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1393.794076] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1393.794221] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1393.794325] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1394.297112] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1394.297368] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1394.297543] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1394.297699] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61970) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1394.298631] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfca5fbc-12e0-4fd5-9689-5a4bebca608e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.307158] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-663254fd-6cea-430c-a214-80d2c8829f53 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.320456] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c3697f9-cbc0-490a-968d-4c36b30ebc6f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.326264] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdfdccda-e2b9-4427-90b4-fdf60aa4e258 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.353277] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180956MB free_disk=150GB free_vcpus=48 pci_devices=None {{(pid=61970) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1394.353420] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1394.353603] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1395.378174] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance d468d634-e0ae-4407-beeb-4c56049ba203 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1395.378485] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 9d77a294-09ca-447c-924b-4feb0124dc70 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1395.378485] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 2b6496dd-2efe-4767-908f-ff695435ef14 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1395.378631] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1395.378772] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1395.422028] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aace89ad-7914-4236-bebe-00a3d8d44fa6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.428724] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-393524b9-4bcb-4f2c-80a0-c27e5a771cd5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.457855] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02cf280c-f080-4863-b617-e38281c5fa5f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.464409] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb85ab8-f621-4fa9-a9be-f7a4e4ec339e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.476861] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1395.996352] env[61970]: ERROR nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [req-9297c846-e010-47dd-b24e-a1f28bcf2952] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fc0f3e4e-5ef5-45e9-8335-cc68743405aa. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-9297c846-e010-47dd-b24e-a1f28bcf2952"}]} [ 1396.011849] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Refreshing inventories for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1396.023538] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Updating ProviderTree inventory for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1396.023706] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1396.033031] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Refreshing aggregate associations for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa, aggregates: None {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1396.048987] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Refreshing trait associations for resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61970) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1396.153733] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65e0bc48-afab-45dc-a523-387fa4e3a902 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.161011] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e2922c3-f053-47d1-be5f-e1e1a07ac03a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.189202] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52dc87f1-400d-4ce2-8148-6591ed1fa9a1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.195816] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be3fc13a-91a2-4f9f-ac40-0a058b00309e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.209675] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1396.739164] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Updated inventory for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with generation 148 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1396.739477] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Updating resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa generation from 148 to 149 during operation: update_inventory {{(pid=61970) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1396.739537] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1396.740924] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1396.741122] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.387s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1397.207496] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1397.207690] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Starting heal instance info cache {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1397.738159] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "refresh_cache-9d77a294-09ca-447c-924b-4feb0124dc70" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1397.738316] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquired lock "refresh_cache-9d77a294-09ca-447c-924b-4feb0124dc70" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1397.738467] env[61970]: DEBUG nova.network.neutron [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Forcefully refreshing network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1398.934029] env[61970]: DEBUG nova.network.neutron [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Updating instance_info_cache with network_info: [{"id": "1e976667-f251-46e2-ae84-6b74c304cf63", "address": "fa:16:3e:cb:7d:1b", "network": {"id": "207c2edd-e6a7-4403-ac4f-0b4cf0e127cc", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1188286473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b739147231ef473288ef6c78eed1caf3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e976667-f2", "ovs_interfaceid": "1e976667-f251-46e2-ae84-6b74c304cf63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1399.436720] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Releasing lock "refresh_cache-9d77a294-09ca-447c-924b-4feb0124dc70" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1399.436936] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Updated the network info_cache for instance {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1399.437164] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1399.437347] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1399.437481] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61970) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1399.437644] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1399.437767] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Cleaning up deleted instances with incomplete migration {{(pid=61970) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1399.940389] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1419.698417] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ed76b881-f70d-4d9a-a6a0-15c6c6c2afd2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "2b6496dd-2efe-4767-908f-ff695435ef14" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1419.698838] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ed76b881-f70d-4d9a-a6a0-15c6c6c2afd2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "2b6496dd-2efe-4767-908f-ff695435ef14" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1420.202073] env[61970]: INFO nova.compute.manager [None req-ed76b881-f70d-4d9a-a6a0-15c6c6c2afd2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Detaching volume 2c5f4217-7dd6-4daf-81b9-d5443ce7ec59 [ 1420.231565] env[61970]: INFO nova.virt.block_device [None req-ed76b881-f70d-4d9a-a6a0-15c6c6c2afd2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Attempting to driver detach volume 2c5f4217-7dd6-4daf-81b9-d5443ce7ec59 from mountpoint /dev/sdb [ 1420.231803] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed76b881-f70d-4d9a-a6a0-15c6c6c2afd2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Volume detach. Driver type: vmdk {{(pid=61970) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1420.231994] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed76b881-f70d-4d9a-a6a0-15c6c6c2afd2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288905', 'volume_id': '2c5f4217-7dd6-4daf-81b9-d5443ce7ec59', 'name': 'volume-2c5f4217-7dd6-4daf-81b9-d5443ce7ec59', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2b6496dd-2efe-4767-908f-ff695435ef14', 'attached_at': '', 'detached_at': '', 'volume_id': '2c5f4217-7dd6-4daf-81b9-d5443ce7ec59', 'serial': '2c5f4217-7dd6-4daf-81b9-d5443ce7ec59'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1420.232913] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f825d50-7df6-40ce-add3-4d35adb91db9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.254596] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-384f5e92-293f-4515-9ba0-b85e6ad8c559 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.260967] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17fcd5a8-d87d-421f-9be4-b1a7a6842b7e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.279545] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e9faac-4cfc-4f6f-9244-190ab2429455 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.294835] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed76b881-f70d-4d9a-a6a0-15c6c6c2afd2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] The volume has not been displaced from its original location: [datastore2] volume-2c5f4217-7dd6-4daf-81b9-d5443ce7ec59/volume-2c5f4217-7dd6-4daf-81b9-d5443ce7ec59.vmdk. No consolidation needed. {{(pid=61970) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1420.299956] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed76b881-f70d-4d9a-a6a0-15c6c6c2afd2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Reconfiguring VM instance instance-0000006e to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1420.300232] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7117384f-046e-4192-9e47-72e4cce0fce7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.317607] env[61970]: DEBUG oslo_vmware.api [None req-ed76b881-f70d-4d9a-a6a0-15c6c6c2afd2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1420.317607] env[61970]: value = "task-1356361" [ 1420.317607] env[61970]: _type = "Task" [ 1420.317607] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1420.324696] env[61970]: DEBUG oslo_vmware.api [None req-ed76b881-f70d-4d9a-a6a0-15c6c6c2afd2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356361, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1420.827171] env[61970]: DEBUG oslo_vmware.api [None req-ed76b881-f70d-4d9a-a6a0-15c6c6c2afd2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356361, 'name': ReconfigVM_Task, 'duration_secs': 0.237979} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1420.827578] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed76b881-f70d-4d9a-a6a0-15c6c6c2afd2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Reconfigured VM instance instance-0000006e to detach disk 2001 {{(pid=61970) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1420.831945] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dff0a83a-d576-4105-b6b4-106d2d30ea39 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.845673] env[61970]: DEBUG oslo_vmware.api [None req-ed76b881-f70d-4d9a-a6a0-15c6c6c2afd2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1420.845673] env[61970]: value = "task-1356362" [ 1420.845673] env[61970]: _type = "Task" [ 1420.845673] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1420.852958] env[61970]: DEBUG oslo_vmware.api [None req-ed76b881-f70d-4d9a-a6a0-15c6c6c2afd2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356362, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1421.355328] env[61970]: DEBUG oslo_vmware.api [None req-ed76b881-f70d-4d9a-a6a0-15c6c6c2afd2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356362, 'name': ReconfigVM_Task, 'duration_secs': 0.12574} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1421.355623] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed76b881-f70d-4d9a-a6a0-15c6c6c2afd2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288905', 'volume_id': '2c5f4217-7dd6-4daf-81b9-d5443ce7ec59', 'name': 'volume-2c5f4217-7dd6-4daf-81b9-d5443ce7ec59', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2b6496dd-2efe-4767-908f-ff695435ef14', 'attached_at': '', 'detached_at': '', 'volume_id': '2c5f4217-7dd6-4daf-81b9-d5443ce7ec59', 'serial': '2c5f4217-7dd6-4daf-81b9-d5443ce7ec59'} {{(pid=61970) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1421.894064] env[61970]: DEBUG nova.objects.instance [None req-ed76b881-f70d-4d9a-a6a0-15c6c6c2afd2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lazy-loading 'flavor' on Instance uuid 2b6496dd-2efe-4767-908f-ff695435ef14 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1422.903025] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ed76b881-f70d-4d9a-a6a0-15c6c6c2afd2 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "2b6496dd-2efe-4767-908f-ff695435ef14" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.204s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1423.997076] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "2b6496dd-2efe-4767-908f-ff695435ef14" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1423.997403] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "2b6496dd-2efe-4767-908f-ff695435ef14" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1423.997574] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "2b6496dd-2efe-4767-908f-ff695435ef14-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1423.997767] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "2b6496dd-2efe-4767-908f-ff695435ef14-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1423.997942] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "2b6496dd-2efe-4767-908f-ff695435ef14-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1424.000129] env[61970]: INFO nova.compute.manager [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Terminating instance [ 1424.001933] env[61970]: DEBUG nova.compute.manager [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1424.002135] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1424.003011] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6020bc91-3848-4065-a882-66c43d6b8cab {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.010804] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1424.011337] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b3b8660b-b5cd-4e8e-9772-336e9b88dba5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.016974] env[61970]: DEBUG oslo_vmware.api [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1424.016974] env[61970]: value = "task-1356363" [ 1424.016974] env[61970]: _type = "Task" [ 1424.016974] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1424.024336] env[61970]: DEBUG oslo_vmware.api [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356363, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1424.527130] env[61970]: DEBUG oslo_vmware.api [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356363, 'name': PowerOffVM_Task, 'duration_secs': 0.178828} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1424.527719] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1424.527719] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1424.527864] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-25566207-1595-4146-aeb8-3238d121d4bb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.592649] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1424.592885] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Deleting contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1424.593085] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Deleting the datastore file [datastore1] 2b6496dd-2efe-4767-908f-ff695435ef14 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1424.593355] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b418e4f0-37e6-4e72-b74f-ad53cb17713a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.600673] env[61970]: DEBUG oslo_vmware.api [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1424.600673] env[61970]: value = "task-1356365" [ 1424.600673] env[61970]: _type = "Task" [ 1424.600673] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1424.607844] env[61970]: DEBUG oslo_vmware.api [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356365, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1425.111265] env[61970]: DEBUG oslo_vmware.api [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356365, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.189786} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1425.111625] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1425.111674] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Deleted contents of the VM from datastore datastore1 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1425.111865] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1425.112056] env[61970]: INFO nova.compute.manager [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1425.112306] env[61970]: DEBUG oslo.service.loopingcall [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1425.112531] env[61970]: DEBUG nova.compute.manager [-] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1425.112627] env[61970]: DEBUG nova.network.neutron [-] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1425.539295] env[61970]: DEBUG nova.compute.manager [req-84ed9aa1-61a3-47bc-a46e-a1d5d6f898d9 req-399f8a34-caa8-48cf-a963-6c255c80ff7c service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Received event network-vif-deleted-7b4d4245-6fe4-4673-aede-924dc3cdce97 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1425.539570] env[61970]: INFO nova.compute.manager [req-84ed9aa1-61a3-47bc-a46e-a1d5d6f898d9 req-399f8a34-caa8-48cf-a963-6c255c80ff7c service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Neutron deleted interface 7b4d4245-6fe4-4673-aede-924dc3cdce97; detaching it from the instance and deleting it from the info cache [ 1425.539805] env[61970]: DEBUG nova.network.neutron [req-84ed9aa1-61a3-47bc-a46e-a1d5d6f898d9 req-399f8a34-caa8-48cf-a963-6c255c80ff7c service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1426.010186] env[61970]: DEBUG nova.network.neutron [-] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1426.042715] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-334313a9-a92e-4552-882a-50af380e9761 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1426.052204] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ace4dbf-b94c-4f4c-ae04-3666ed72e824 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1426.076835] env[61970]: DEBUG nova.compute.manager [req-84ed9aa1-61a3-47bc-a46e-a1d5d6f898d9 req-399f8a34-caa8-48cf-a963-6c255c80ff7c service nova] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Detach interface failed, port_id=7b4d4245-6fe4-4673-aede-924dc3cdce97, reason: Instance 2b6496dd-2efe-4767-908f-ff695435ef14 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1426.513583] env[61970]: INFO nova.compute.manager [-] [instance: 2b6496dd-2efe-4767-908f-ff695435ef14] Took 1.40 seconds to deallocate network for instance. [ 1427.020409] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1427.020675] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1427.020902] env[61970]: DEBUG nova.objects.instance [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lazy-loading 'resources' on Instance uuid 2b6496dd-2efe-4767-908f-ff695435ef14 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1427.572704] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afbc1399-2a55-46da-9c27-172ab0792530 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.579938] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2641d79-3e12-4d92-b75a-27a9b469fb70 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.608402] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf813d7-dc6e-4a2a-b6d7-92828a58301c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.615044] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d25390-f1d3-45dc-b3c1-08a1a0146515 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.627504] env[61970]: DEBUG nova.compute.provider_tree [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1428.156429] env[61970]: DEBUG nova.scheduler.client.report [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Updated inventory for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with generation 149 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1428.156704] env[61970]: DEBUG nova.compute.provider_tree [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Updating resource provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa generation from 149 to 150 during operation: update_inventory {{(pid=61970) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1428.156889] env[61970]: DEBUG nova.compute.provider_tree [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Updating inventory in ProviderTree for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1428.662310] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.641s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1428.682700] env[61970]: INFO nova.scheduler.client.report [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Deleted allocations for instance 2b6496dd-2efe-4767-908f-ff695435ef14 [ 1429.190826] env[61970]: DEBUG oslo_concurrency.lockutils [None req-ed551f0b-7141-40c7-b419-134032fbb1be tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "2b6496dd-2efe-4767-908f-ff695435ef14" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.193s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1434.103061] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "9d77a294-09ca-447c-924b-4feb0124dc70" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1434.103061] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "9d77a294-09ca-447c-924b-4feb0124dc70" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1434.103061] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "9d77a294-09ca-447c-924b-4feb0124dc70-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1434.103061] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "9d77a294-09ca-447c-924b-4feb0124dc70-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1434.103601] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "9d77a294-09ca-447c-924b-4feb0124dc70-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1434.105509] env[61970]: INFO nova.compute.manager [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Terminating instance [ 1434.107239] env[61970]: DEBUG nova.compute.manager [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1434.107434] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1434.108274] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f9713e6-30c3-4dab-9f1e-11e51c9d042e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1434.115931] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1434.116176] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e78171be-83e1-4fa8-ad97-a28e8d040ec6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1434.123404] env[61970]: DEBUG oslo_vmware.api [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1434.123404] env[61970]: value = "task-1356367" [ 1434.123404] env[61970]: _type = "Task" [ 1434.123404] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1434.130445] env[61970]: DEBUG oslo_vmware.api [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356367, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1434.633272] env[61970]: DEBUG oslo_vmware.api [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356367, 'name': PowerOffVM_Task, 'duration_secs': 0.190029} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1434.633553] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1434.633735] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1434.633984] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f127bca-118e-4e03-b77e-483bff58b6cd {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1434.699123] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1434.699385] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1434.699572] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Deleting the datastore file [datastore2] 9d77a294-09ca-447c-924b-4feb0124dc70 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1434.699876] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d541c75e-f915-48ca-922a-26cdd1724bcf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1434.706494] env[61970]: DEBUG oslo_vmware.api [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1434.706494] env[61970]: value = "task-1356369" [ 1434.706494] env[61970]: _type = "Task" [ 1434.706494] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1434.714053] env[61970]: DEBUG oslo_vmware.api [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356369, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1435.216396] env[61970]: DEBUG oslo_vmware.api [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356369, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177409} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1435.216799] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1435.216862] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1435.217020] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1435.217219] env[61970]: INFO nova.compute.manager [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1435.217464] env[61970]: DEBUG oslo.service.loopingcall [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1435.217667] env[61970]: DEBUG nova.compute.manager [-] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1435.217767] env[61970]: DEBUG nova.network.neutron [-] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1435.459178] env[61970]: DEBUG nova.compute.manager [req-0529e11d-ca54-4b1e-a714-1ed5792d13ac req-9932be78-154c-4e67-8501-d6d9e87f315f service nova] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Received event network-vif-deleted-1e976667-f251-46e2-ae84-6b74c304cf63 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1435.459408] env[61970]: INFO nova.compute.manager [req-0529e11d-ca54-4b1e-a714-1ed5792d13ac req-9932be78-154c-4e67-8501-d6d9e87f315f service nova] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Neutron deleted interface 1e976667-f251-46e2-ae84-6b74c304cf63; detaching it from the instance and deleting it from the info cache [ 1435.459585] env[61970]: DEBUG nova.network.neutron [req-0529e11d-ca54-4b1e-a714-1ed5792d13ac req-9932be78-154c-4e67-8501-d6d9e87f315f service nova] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1435.940022] env[61970]: DEBUG nova.network.neutron [-] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1435.962210] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-888afab9-c24a-4c54-a8cd-4ebfb5524f8b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1435.970704] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3380d2e-b474-4cdb-8e6a-182743393559 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1435.993928] env[61970]: DEBUG nova.compute.manager [req-0529e11d-ca54-4b1e-a714-1ed5792d13ac req-9932be78-154c-4e67-8501-d6d9e87f315f service nova] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Detach interface failed, port_id=1e976667-f251-46e2-ae84-6b74c304cf63, reason: Instance 9d77a294-09ca-447c-924b-4feb0124dc70 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1436.442119] env[61970]: INFO nova.compute.manager [-] [instance: 9d77a294-09ca-447c-924b-4feb0124dc70] Took 1.22 seconds to deallocate network for instance. [ 1436.948465] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1436.948723] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1436.948951] env[61970]: DEBUG nova.objects.instance [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lazy-loading 'resources' on Instance uuid 9d77a294-09ca-447c-924b-4feb0124dc70 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1437.493360] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbae2b90-0c09-4052-9bb7-305a4281d21e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1437.500691] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79ee0973-1b42-41a6-a80a-d8577b82e8b9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1437.529722] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3714f2cc-dc40-4928-b76e-bc447babee87 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1437.536345] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d0f744-0aa7-4fcd-9696-00540e01402a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1437.548773] env[61970]: DEBUG nova.compute.provider_tree [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1438.051731] env[61970]: DEBUG nova.scheduler.client.report [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1438.557426] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.608s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1438.580021] env[61970]: INFO nova.scheduler.client.report [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Deleted allocations for instance 9d77a294-09ca-447c-924b-4feb0124dc70 [ 1439.089476] env[61970]: DEBUG oslo_concurrency.lockutils [None req-d7038f8b-fc20-4cb7-9ccd-4e74b38bf163 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "9d77a294-09ca-447c-924b-4feb0124dc70" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.987s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1439.659408] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "d468d634-e0ae-4407-beeb-4c56049ba203" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1439.659791] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "d468d634-e0ae-4407-beeb-4c56049ba203" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1439.660102] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "d468d634-e0ae-4407-beeb-4c56049ba203-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1439.660314] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "d468d634-e0ae-4407-beeb-4c56049ba203-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1439.660570] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "d468d634-e0ae-4407-beeb-4c56049ba203-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1439.662881] env[61970]: INFO nova.compute.manager [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Terminating instance [ 1439.664745] env[61970]: DEBUG nova.compute.manager [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1439.664990] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1439.665930] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc509a3-2615-4fbf-8368-cc5c9885eb76 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.673532] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1439.674094] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b53a3ad8-0d89-4ac1-8f02-5e1cd04b8740 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.680515] env[61970]: DEBUG oslo_vmware.api [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1439.680515] env[61970]: value = "task-1356370" [ 1439.680515] env[61970]: _type = "Task" [ 1439.680515] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1439.688473] env[61970]: DEBUG oslo_vmware.api [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356370, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1440.190062] env[61970]: DEBUG oslo_vmware.api [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356370, 'name': PowerOffVM_Task, 'duration_secs': 0.19077} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1440.190377] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1440.190516] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1440.190781] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3037d7b8-22be-4516-a0ef-3de23956fefe {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.255420] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1440.255651] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1440.255839] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Deleting the datastore file [datastore2] d468d634-e0ae-4407-beeb-4c56049ba203 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1440.256138] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4104ef81-1dec-4cd1-9d54-2b49fedded38 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.262268] env[61970]: DEBUG oslo_vmware.api [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for the task: (returnval){ [ 1440.262268] env[61970]: value = "task-1356372" [ 1440.262268] env[61970]: _type = "Task" [ 1440.262268] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1440.269547] env[61970]: DEBUG oslo_vmware.api [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356372, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1440.771996] env[61970]: DEBUG oslo_vmware.api [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Task: {'id': task-1356372, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13936} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1440.772410] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1440.772457] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1440.772621] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1440.772798] env[61970]: INFO nova.compute.manager [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1440.773106] env[61970]: DEBUG oslo.service.loopingcall [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1440.773312] env[61970]: DEBUG nova.compute.manager [-] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1440.773408] env[61970]: DEBUG nova.network.neutron [-] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1441.012232] env[61970]: DEBUG nova.compute.manager [req-14008276-5c4c-417e-b746-8cb2fb92a39e req-f3210c14-b302-468e-ae21-3deeb8b0d156 service nova] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Received event network-vif-deleted-1a5feac9-a368-4804-bb4a-648e682d6682 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1441.012445] env[61970]: INFO nova.compute.manager [req-14008276-5c4c-417e-b746-8cb2fb92a39e req-f3210c14-b302-468e-ae21-3deeb8b0d156 service nova] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Neutron deleted interface 1a5feac9-a368-4804-bb4a-648e682d6682; detaching it from the instance and deleting it from the info cache [ 1441.012627] env[61970]: DEBUG nova.network.neutron [req-14008276-5c4c-417e-b746-8cb2fb92a39e req-f3210c14-b302-468e-ae21-3deeb8b0d156 service nova] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1441.488811] env[61970]: DEBUG nova.network.neutron [-] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1441.515848] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1340c183-1f00-49cd-b52d-c73c3448bf12 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.526917] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfc1d180-c6b8-4bc2-b7a3-315925629656 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.548862] env[61970]: DEBUG nova.compute.manager [req-14008276-5c4c-417e-b746-8cb2fb92a39e req-f3210c14-b302-468e-ae21-3deeb8b0d156 service nova] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Detach interface failed, port_id=1a5feac9-a368-4804-bb4a-648e682d6682, reason: Instance d468d634-e0ae-4407-beeb-4c56049ba203 could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1441.992259] env[61970]: INFO nova.compute.manager [-] [instance: d468d634-e0ae-4407-beeb-4c56049ba203] Took 1.22 seconds to deallocate network for instance. [ 1442.499068] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1442.499367] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1442.499592] env[61970]: DEBUG nova.objects.instance [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lazy-loading 'resources' on Instance uuid d468d634-e0ae-4407-beeb-4c56049ba203 {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1443.032807] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e26a756-dbac-45c7-a782-75d9e7bc6cef {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1443.040093] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7cb9b82-8784-47fc-b9ea-8e13f0aefcd6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1443.068611] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef4c45b-147e-4365-a63c-17a3e9865bd3 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1443.075215] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e79c5d-ae66-4f56-b303-f4ad921b03eb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1443.088213] env[61970]: DEBUG nova.compute.provider_tree [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1443.591767] env[61970]: DEBUG nova.scheduler.client.report [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1444.097402] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.598s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1444.117598] env[61970]: INFO nova.scheduler.client.report [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Deleted allocations for instance d468d634-e0ae-4407-beeb-4c56049ba203 [ 1444.627451] env[61970]: DEBUG oslo_concurrency.lockutils [None req-4807b32b-0eed-493a-94e8-a25de8540fd1 tempest-ServerRescueNegativeTestJSON-297037660 tempest-ServerRescueNegativeTestJSON-297037660-project-member] Lock "d468d634-e0ae-4407-beeb-4c56049ba203" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.968s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1446.264747] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1448.860855] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1448.861233] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1449.363872] env[61970]: DEBUG nova.compute.manager [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1449.885918] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1449.886233] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1449.887625] env[61970]: INFO nova.compute.claims [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1450.260123] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1450.925308] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37244a38-c21d-4c2b-9b4a-fc162110b834 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1450.932722] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3358953b-970f-4204-9e0c-715f85f5565b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1450.966640] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af425ead-6e02-405a-afde-6484289148f2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1450.974198] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9c90b3-0f2f-411d-8a2d-f77813d52d71 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1450.988137] env[61970]: DEBUG nova.compute.provider_tree [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1451.259858] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1451.491453] env[61970]: DEBUG nova.scheduler.client.report [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1451.762819] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1451.996595] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.110s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1451.997158] env[61970]: DEBUG nova.compute.manager [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1451.999671] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.237s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1452.000417] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1452.000417] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61970) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1452.001053] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1243d9d-2d36-467f-97c1-0a463fca63da {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1452.009406] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fefbeaaa-a82b-4377-a4e0-5c210c1d1261 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1452.023503] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c8169b7-a7a4-41c6-a763-654379cc3257 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1452.030581] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1714d56-81bf-4eda-8433-a30c016cb148 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1452.061157] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180956MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61970) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1452.061297] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1452.061484] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1452.503063] env[61970]: DEBUG nova.compute.utils [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1452.505192] env[61970]: DEBUG nova.compute.manager [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1452.505308] env[61970]: DEBUG nova.network.neutron [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1452.546691] env[61970]: DEBUG nova.policy [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e5e0b3445e8b45ad940e894e6db26165', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '92a836a268414a49a13fd816aa1c2407', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 1452.779984] env[61970]: DEBUG nova.network.neutron [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Successfully created port: 01289152-eff0-4751-94e5-732eaefc7cb0 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1453.008246] env[61970]: DEBUG nova.compute.manager [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1453.079935] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 99e6c9f1-6ecd-439b-a716-c378c18ec9bb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1453.080166] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1453.080312] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1453.103756] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e652c9-19cd-4414-bca7-dbb746b77155 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1453.111580] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f03f42a2-2835-40cd-af55-fd61dc34fe20 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1453.140875] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-660c2c01-e7d9-49d8-a8cc-fb121dcae735 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1453.147738] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a903111d-aad7-497c-92fa-a9fb1a399b71 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1453.160175] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1453.663140] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1454.019106] env[61970]: DEBUG nova.compute.manager [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1454.043956] env[61970]: DEBUG nova.virt.hardware [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1454.044383] env[61970]: DEBUG nova.virt.hardware [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1454.044588] env[61970]: DEBUG nova.virt.hardware [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1454.044782] env[61970]: DEBUG nova.virt.hardware [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1454.044933] env[61970]: DEBUG nova.virt.hardware [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1454.045100] env[61970]: DEBUG nova.virt.hardware [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1454.045322] env[61970]: DEBUG nova.virt.hardware [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1454.045484] env[61970]: DEBUG nova.virt.hardware [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1454.045652] env[61970]: DEBUG nova.virt.hardware [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1454.045820] env[61970]: DEBUG nova.virt.hardware [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1454.045995] env[61970]: DEBUG nova.virt.hardware [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1454.046900] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07873e36-569c-4cc7-a111-d994fa02e7ae {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1454.055085] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a214a0-fb61-4ca3-a5a5-81e3e73301be {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1454.167927] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1454.168303] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.107s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1454.195135] env[61970]: DEBUG nova.compute.manager [req-3ba5d550-cc7f-4b4b-9487-a527aa491d24 req-44069582-5e2a-4943-aaf8-848846967089 service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Received event network-vif-plugged-01289152-eff0-4751-94e5-732eaefc7cb0 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1454.195362] env[61970]: DEBUG oslo_concurrency.lockutils [req-3ba5d550-cc7f-4b4b-9487-a527aa491d24 req-44069582-5e2a-4943-aaf8-848846967089 service nova] Acquiring lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1454.195570] env[61970]: DEBUG oslo_concurrency.lockutils [req-3ba5d550-cc7f-4b4b-9487-a527aa491d24 req-44069582-5e2a-4943-aaf8-848846967089 service nova] Lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1454.195743] env[61970]: DEBUG oslo_concurrency.lockutils [req-3ba5d550-cc7f-4b4b-9487-a527aa491d24 req-44069582-5e2a-4943-aaf8-848846967089 service nova] Lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1454.195952] env[61970]: DEBUG nova.compute.manager [req-3ba5d550-cc7f-4b4b-9487-a527aa491d24 req-44069582-5e2a-4943-aaf8-848846967089 service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] No waiting events found dispatching network-vif-plugged-01289152-eff0-4751-94e5-732eaefc7cb0 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1454.196097] env[61970]: WARNING nova.compute.manager [req-3ba5d550-cc7f-4b4b-9487-a527aa491d24 req-44069582-5e2a-4943-aaf8-848846967089 service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Received unexpected event network-vif-plugged-01289152-eff0-4751-94e5-732eaefc7cb0 for instance with vm_state building and task_state spawning. [ 1454.247025] env[61970]: DEBUG nova.network.neutron [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Successfully updated port: 01289152-eff0-4751-94e5-732eaefc7cb0 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1454.751531] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1454.751676] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquired lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1454.751830] env[61970]: DEBUG nova.network.neutron [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1455.163959] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1455.164351] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1455.164437] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1455.259989] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1455.260072] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61970) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1455.282992] env[61970]: DEBUG nova.network.neutron [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1455.399950] env[61970]: DEBUG nova.network.neutron [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Updating instance_info_cache with network_info: [{"id": "01289152-eff0-4751-94e5-732eaefc7cb0", "address": "fa:16:3e:d9:0e:c3", "network": {"id": "8c857731-1df4-4254-93c3-5a86f4a2dd2b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1262016675-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "92a836a268414a49a13fd816aa1c2407", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01289152-ef", "ovs_interfaceid": "01289152-eff0-4751-94e5-732eaefc7cb0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1455.902091] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Releasing lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1455.902385] env[61970]: DEBUG nova.compute.manager [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Instance network_info: |[{"id": "01289152-eff0-4751-94e5-732eaefc7cb0", "address": "fa:16:3e:d9:0e:c3", "network": {"id": "8c857731-1df4-4254-93c3-5a86f4a2dd2b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1262016675-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "92a836a268414a49a13fd816aa1c2407", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01289152-ef", "ovs_interfaceid": "01289152-eff0-4751-94e5-732eaefc7cb0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1455.902806] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d9:0e:c3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9c621a9c-66f5-426a-8aab-bd8b2e912106', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '01289152-eff0-4751-94e5-732eaefc7cb0', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1455.910254] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Creating folder: Project (92a836a268414a49a13fd816aa1c2407). Parent ref: group-v288740. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1455.910689] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a88bf25d-12d7-423f-8633-53856d5eb3d9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1455.921714] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Created folder: Project (92a836a268414a49a13fd816aa1c2407) in parent group-v288740. [ 1455.921899] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Creating folder: Instances. Parent ref: group-v288906. {{(pid=61970) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1455.922133] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d4bb5f48-cdaa-4545-8f42-d8403031e9a5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1455.930814] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Created folder: Instances in parent group-v288906. [ 1455.931050] env[61970]: DEBUG oslo.service.loopingcall [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1455.931255] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1455.931451] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e8a6f4df-f6df-4551-a5ff-9225db33fe56 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1455.949692] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1455.949692] env[61970]: value = "task-1356375" [ 1455.949692] env[61970]: _type = "Task" [ 1455.949692] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1455.956747] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356375, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1456.221666] env[61970]: DEBUG nova.compute.manager [req-d3490901-4c0c-42da-9e6b-87a3314fe206 req-d444705c-0ac3-4282-a7b9-d74f5f03acb7 service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Received event network-changed-01289152-eff0-4751-94e5-732eaefc7cb0 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1456.221916] env[61970]: DEBUG nova.compute.manager [req-d3490901-4c0c-42da-9e6b-87a3314fe206 req-d444705c-0ac3-4282-a7b9-d74f5f03acb7 service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Refreshing instance network info cache due to event network-changed-01289152-eff0-4751-94e5-732eaefc7cb0. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1456.222031] env[61970]: DEBUG oslo_concurrency.lockutils [req-d3490901-4c0c-42da-9e6b-87a3314fe206 req-d444705c-0ac3-4282-a7b9-d74f5f03acb7 service nova] Acquiring lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1456.222149] env[61970]: DEBUG oslo_concurrency.lockutils [req-d3490901-4c0c-42da-9e6b-87a3314fe206 req-d444705c-0ac3-4282-a7b9-d74f5f03acb7 service nova] Acquired lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1456.222316] env[61970]: DEBUG nova.network.neutron [req-d3490901-4c0c-42da-9e6b-87a3314fe206 req-d444705c-0ac3-4282-a7b9-d74f5f03acb7 service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Refreshing network info cache for port 01289152-eff0-4751-94e5-732eaefc7cb0 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1456.260286] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1456.260448] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Starting heal instance info cache {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1456.459074] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356375, 'name': CreateVM_Task, 'duration_secs': 0.318804} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1456.459253] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1456.466069] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1456.466257] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1456.466625] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1456.466876] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ae07386-4a74-45fc-80d0-b30921520601 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1456.471407] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1456.471407] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52915c80-9c5b-c0b4-52bb-4e410c6bbde4" [ 1456.471407] env[61970]: _type = "Task" [ 1456.471407] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1456.478919] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52915c80-9c5b-c0b4-52bb-4e410c6bbde4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1456.762663] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Didn't find any instances for network info cache update. {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1456.981537] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52915c80-9c5b-c0b4-52bb-4e410c6bbde4, 'name': SearchDatastore_Task, 'duration_secs': 0.00911} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1456.981848] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1456.982102] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1456.982349] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1456.982505] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1456.982714] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1456.982965] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-663ca2ab-582f-4401-8bc6-b2854dd68ee8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1456.990948] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1456.991151] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1456.991862] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67a3785d-e8d1-4c17-8486-3eadfb6978f7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1456.996492] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1456.996492] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d7d785-e57d-b04d-b069-af9ffc3ff472" [ 1456.996492] env[61970]: _type = "Task" [ 1456.996492] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1457.003619] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d7d785-e57d-b04d-b069-af9ffc3ff472, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1457.125530] env[61970]: DEBUG nova.network.neutron [req-d3490901-4c0c-42da-9e6b-87a3314fe206 req-d444705c-0ac3-4282-a7b9-d74f5f03acb7 service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Updated VIF entry in instance network info cache for port 01289152-eff0-4751-94e5-732eaefc7cb0. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1457.125887] env[61970]: DEBUG nova.network.neutron [req-d3490901-4c0c-42da-9e6b-87a3314fe206 req-d444705c-0ac3-4282-a7b9-d74f5f03acb7 service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Updating instance_info_cache with network_info: [{"id": "01289152-eff0-4751-94e5-732eaefc7cb0", "address": "fa:16:3e:d9:0e:c3", "network": {"id": "8c857731-1df4-4254-93c3-5a86f4a2dd2b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1262016675-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "92a836a268414a49a13fd816aa1c2407", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01289152-ef", "ovs_interfaceid": "01289152-eff0-4751-94e5-732eaefc7cb0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1457.507342] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52d7d785-e57d-b04d-b069-af9ffc3ff472, 'name': SearchDatastore_Task, 'duration_secs': 0.007815} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1457.508107] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2884285-9c5a-4e25-a0b1-d818d0e26243 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1457.513141] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1457.513141] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ce7c05-a722-8365-1305-af3bb66849a4" [ 1457.513141] env[61970]: _type = "Task" [ 1457.513141] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1457.520431] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ce7c05-a722-8365-1305-af3bb66849a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1457.628360] env[61970]: DEBUG oslo_concurrency.lockutils [req-d3490901-4c0c-42da-9e6b-87a3314fe206 req-d444705c-0ac3-4282-a7b9-d74f5f03acb7 service nova] Releasing lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1457.757736] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1458.023783] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52ce7c05-a722-8365-1305-af3bb66849a4, 'name': SearchDatastore_Task, 'duration_secs': 0.008992} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1458.024064] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1458.024354] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 99e6c9f1-6ecd-439b-a716-c378c18ec9bb/99e6c9f1-6ecd-439b-a716-c378c18ec9bb.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1458.024625] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-69633de2-f039-46af-8a8a-d7b682e2dfce {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1458.030586] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1458.030586] env[61970]: value = "task-1356376" [ 1458.030586] env[61970]: _type = "Task" [ 1458.030586] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1458.037791] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356376, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1458.263016] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1458.541059] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356376, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.436641} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1458.541059] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 99e6c9f1-6ecd-439b-a716-c378c18ec9bb/99e6c9f1-6ecd-439b-a716-c378c18ec9bb.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1458.541507] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1458.541622] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4642b20b-c410-4cf8-9551-77a3978135f2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1458.547276] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1458.547276] env[61970]: value = "task-1356377" [ 1458.547276] env[61970]: _type = "Task" [ 1458.547276] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1458.554831] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356377, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1459.056177] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356377, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.147242} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1459.056448] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1459.057208] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f8735e-3122-47b7-abdd-9617ce7dad16 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1459.079539] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 99e6c9f1-6ecd-439b-a716-c378c18ec9bb/99e6c9f1-6ecd-439b-a716-c378c18ec9bb.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1459.079770] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f286505e-cf55-4776-adde-b9e4da0b88d2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1459.097952] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1459.097952] env[61970]: value = "task-1356378" [ 1459.097952] env[61970]: _type = "Task" [ 1459.097952] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1459.105095] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356378, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1459.607644] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356378, 'name': ReconfigVM_Task, 'duration_secs': 0.2913} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1459.608090] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 99e6c9f1-6ecd-439b-a716-c378c18ec9bb/99e6c9f1-6ecd-439b-a716-c378c18ec9bb.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1459.608559] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a318646c-a2bb-43ec-9726-a525b82a3b5d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1459.614644] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1459.614644] env[61970]: value = "task-1356379" [ 1459.614644] env[61970]: _type = "Task" [ 1459.614644] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1459.623200] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356379, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1460.124092] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356379, 'name': Rename_Task, 'duration_secs': 0.131806} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1460.124355] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1460.124597] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-63a5aad7-a201-4ccc-8a16-6ef358a19085 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1460.130511] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1460.130511] env[61970]: value = "task-1356380" [ 1460.130511] env[61970]: _type = "Task" [ 1460.130511] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1460.137536] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356380, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1460.640457] env[61970]: DEBUG oslo_vmware.api [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356380, 'name': PowerOnVM_Task, 'duration_secs': 0.434433} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1460.640840] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1460.640895] env[61970]: INFO nova.compute.manager [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Took 6.62 seconds to spawn the instance on the hypervisor. [ 1460.641090] env[61970]: DEBUG nova.compute.manager [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1460.641841] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d3132d-7958-41c9-8c75-9d12a634108f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1461.158641] env[61970]: INFO nova.compute.manager [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Took 11.29 seconds to build instance. [ 1461.660193] env[61970]: DEBUG oslo_concurrency.lockutils [None req-df8999a9-9320-4809-9339-13c2d1c43ca8 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.799s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1462.343849] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "39117eed-03b2-4619-b8d6-04f09309a7ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1462.344102] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "39117eed-03b2-4619-b8d6-04f09309a7ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1462.846533] env[61970]: DEBUG nova.compute.manager [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Starting instance... {{(pid=61970) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1463.369786] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1463.370010] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1463.371483] env[61970]: INFO nova.compute.claims [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1464.415431] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf653b0a-094f-4c51-bf95-71bee626fe3f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1464.422511] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e4e376-31a3-4232-9d67-8b2653ece2d6 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1464.450984] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edab0565-c509-4424-a6e6-f4198940a8c1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1464.457390] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7888f8a0-3317-4d6c-b4c2-d49b47f1c128 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1464.469561] env[61970]: DEBUG nova.compute.provider_tree [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1464.972745] env[61970]: DEBUG nova.scheduler.client.report [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1465.478691] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.108s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1465.479288] env[61970]: DEBUG nova.compute.manager [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Start building networks asynchronously for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1465.985049] env[61970]: DEBUG nova.compute.utils [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Using /dev/sd instead of None {{(pid=61970) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1465.985921] env[61970]: DEBUG nova.compute.manager [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Allocating IP information in the background. {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1465.986542] env[61970]: DEBUG nova.network.neutron [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] allocate_for_instance() {{(pid=61970) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1466.090182] env[61970]: DEBUG nova.policy [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e5e0b3445e8b45ad940e894e6db26165', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '92a836a268414a49a13fd816aa1c2407', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61970) authorize /opt/stack/nova/nova/policy.py:201}} [ 1466.334269] env[61970]: DEBUG nova.network.neutron [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Successfully created port: 27454272-523e-44c5-8c57-dd33bdd14c61 {{(pid=61970) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1466.489716] env[61970]: DEBUG nova.compute.manager [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Start building block device mappings for instance. {{(pid=61970) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1467.500990] env[61970]: DEBUG nova.compute.manager [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Start spawning the instance on the hypervisor. {{(pid=61970) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1467.527304] env[61970]: DEBUG nova.virt.hardware [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T13:36:04Z,direct_url=,disk_format='vmdk',id=7ff5af44-3f73-4c84-82e5-a1543db26d03,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='151f02d8f9934c8489ca452ac8127e81',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T13:36:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1467.527562] env[61970]: DEBUG nova.virt.hardware [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1467.527760] env[61970]: DEBUG nova.virt.hardware [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1467.527975] env[61970]: DEBUG nova.virt.hardware [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1467.528143] env[61970]: DEBUG nova.virt.hardware [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1467.528297] env[61970]: DEBUG nova.virt.hardware [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1467.528509] env[61970]: DEBUG nova.virt.hardware [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1467.528698] env[61970]: DEBUG nova.virt.hardware [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1467.528885] env[61970]: DEBUG nova.virt.hardware [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1467.529067] env[61970]: DEBUG nova.virt.hardware [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1467.529250] env[61970]: DEBUG nova.virt.hardware [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1467.530142] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13f67fde-7d4a-46f1-9714-43e2c9d2bf8d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1467.538149] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d59f37-7b4a-4872-bd69-0855d4da6511 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1467.702771] env[61970]: DEBUG nova.compute.manager [req-534b9494-cc67-49c3-9c8b-f5eb7d04b196 req-e82369ef-2b02-4b4c-a29f-46df573e5065 service nova] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Received event network-vif-plugged-27454272-523e-44c5-8c57-dd33bdd14c61 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1467.702998] env[61970]: DEBUG oslo_concurrency.lockutils [req-534b9494-cc67-49c3-9c8b-f5eb7d04b196 req-e82369ef-2b02-4b4c-a29f-46df573e5065 service nova] Acquiring lock "39117eed-03b2-4619-b8d6-04f09309a7ea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1467.703232] env[61970]: DEBUG oslo_concurrency.lockutils [req-534b9494-cc67-49c3-9c8b-f5eb7d04b196 req-e82369ef-2b02-4b4c-a29f-46df573e5065 service nova] Lock "39117eed-03b2-4619-b8d6-04f09309a7ea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1467.703405] env[61970]: DEBUG oslo_concurrency.lockutils [req-534b9494-cc67-49c3-9c8b-f5eb7d04b196 req-e82369ef-2b02-4b4c-a29f-46df573e5065 service nova] Lock "39117eed-03b2-4619-b8d6-04f09309a7ea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1467.703621] env[61970]: DEBUG nova.compute.manager [req-534b9494-cc67-49c3-9c8b-f5eb7d04b196 req-e82369ef-2b02-4b4c-a29f-46df573e5065 service nova] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] No waiting events found dispatching network-vif-plugged-27454272-523e-44c5-8c57-dd33bdd14c61 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1467.703821] env[61970]: WARNING nova.compute.manager [req-534b9494-cc67-49c3-9c8b-f5eb7d04b196 req-e82369ef-2b02-4b4c-a29f-46df573e5065 service nova] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Received unexpected event network-vif-plugged-27454272-523e-44c5-8c57-dd33bdd14c61 for instance with vm_state building and task_state spawning. [ 1467.789473] env[61970]: DEBUG nova.network.neutron [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Successfully updated port: 27454272-523e-44c5-8c57-dd33bdd14c61 {{(pid=61970) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1468.292155] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "refresh_cache-39117eed-03b2-4619-b8d6-04f09309a7ea" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1468.292358] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquired lock "refresh_cache-39117eed-03b2-4619-b8d6-04f09309a7ea" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1468.292437] env[61970]: DEBUG nova.network.neutron [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1468.829341] env[61970]: DEBUG nova.network.neutron [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Instance cache missing network info. {{(pid=61970) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1468.963427] env[61970]: DEBUG nova.network.neutron [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Updating instance_info_cache with network_info: [{"id": "27454272-523e-44c5-8c57-dd33bdd14c61", "address": "fa:16:3e:f2:29:1e", "network": {"id": "8c857731-1df4-4254-93c3-5a86f4a2dd2b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1262016675-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "92a836a268414a49a13fd816aa1c2407", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27454272-52", "ovs_interfaceid": "27454272-523e-44c5-8c57-dd33bdd14c61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1469.466716] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Releasing lock "refresh_cache-39117eed-03b2-4619-b8d6-04f09309a7ea" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1469.467070] env[61970]: DEBUG nova.compute.manager [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Instance network_info: |[{"id": "27454272-523e-44c5-8c57-dd33bdd14c61", "address": "fa:16:3e:f2:29:1e", "network": {"id": "8c857731-1df4-4254-93c3-5a86f4a2dd2b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1262016675-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "92a836a268414a49a13fd816aa1c2407", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27454272-52", "ovs_interfaceid": "27454272-523e-44c5-8c57-dd33bdd14c61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61970) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1469.467522] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f2:29:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9c621a9c-66f5-426a-8aab-bd8b2e912106', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '27454272-523e-44c5-8c57-dd33bdd14c61', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1469.475117] env[61970]: DEBUG oslo.service.loopingcall [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1469.475331] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1469.475559] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d8917bbd-743f-40c2-8b97-1f2ad224afaa {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1469.496625] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1469.496625] env[61970]: value = "task-1356381" [ 1469.496625] env[61970]: _type = "Task" [ 1469.496625] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1469.504276] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356381, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1469.734091] env[61970]: DEBUG nova.compute.manager [req-e8cc27bb-d848-4f34-8d98-a453dac24d2f req-f48d6add-b976-4f23-a541-bbda735847a5 service nova] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Received event network-changed-27454272-523e-44c5-8c57-dd33bdd14c61 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1469.734399] env[61970]: DEBUG nova.compute.manager [req-e8cc27bb-d848-4f34-8d98-a453dac24d2f req-f48d6add-b976-4f23-a541-bbda735847a5 service nova] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Refreshing instance network info cache due to event network-changed-27454272-523e-44c5-8c57-dd33bdd14c61. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1469.734717] env[61970]: DEBUG oslo_concurrency.lockutils [req-e8cc27bb-d848-4f34-8d98-a453dac24d2f req-f48d6add-b976-4f23-a541-bbda735847a5 service nova] Acquiring lock "refresh_cache-39117eed-03b2-4619-b8d6-04f09309a7ea" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1469.734952] env[61970]: DEBUG oslo_concurrency.lockutils [req-e8cc27bb-d848-4f34-8d98-a453dac24d2f req-f48d6add-b976-4f23-a541-bbda735847a5 service nova] Acquired lock "refresh_cache-39117eed-03b2-4619-b8d6-04f09309a7ea" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1469.735228] env[61970]: DEBUG nova.network.neutron [req-e8cc27bb-d848-4f34-8d98-a453dac24d2f req-f48d6add-b976-4f23-a541-bbda735847a5 service nova] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Refreshing network info cache for port 27454272-523e-44c5-8c57-dd33bdd14c61 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1470.005920] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356381, 'name': CreateVM_Task, 'duration_secs': 0.395675} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1470.006304] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1470.006753] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1470.006926] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1470.007295] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1470.007526] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a8b6c23-30bf-4367-9581-6ace07e11be8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1470.011713] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1470.011713] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5264ef7d-459d-27c2-8139-1e62cc3be0bb" [ 1470.011713] env[61970]: _type = "Task" [ 1470.011713] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1470.018995] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5264ef7d-459d-27c2-8139-1e62cc3be0bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1470.424870] env[61970]: DEBUG nova.network.neutron [req-e8cc27bb-d848-4f34-8d98-a453dac24d2f req-f48d6add-b976-4f23-a541-bbda735847a5 service nova] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Updated VIF entry in instance network info cache for port 27454272-523e-44c5-8c57-dd33bdd14c61. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1470.425293] env[61970]: DEBUG nova.network.neutron [req-e8cc27bb-d848-4f34-8d98-a453dac24d2f req-f48d6add-b976-4f23-a541-bbda735847a5 service nova] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Updating instance_info_cache with network_info: [{"id": "27454272-523e-44c5-8c57-dd33bdd14c61", "address": "fa:16:3e:f2:29:1e", "network": {"id": "8c857731-1df4-4254-93c3-5a86f4a2dd2b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1262016675-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "92a836a268414a49a13fd816aa1c2407", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27454272-52", "ovs_interfaceid": "27454272-523e-44c5-8c57-dd33bdd14c61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1470.522785] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5264ef7d-459d-27c2-8139-1e62cc3be0bb, 'name': SearchDatastore_Task, 'duration_secs': 0.009302} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1470.523042] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1470.523290] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Processing image 7ff5af44-3f73-4c84-82e5-a1543db26d03 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1470.523523] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1470.523703] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1470.523942] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1470.524470] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6e7a3a41-0387-41e7-964a-c6d9574053c4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1470.532325] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1470.532499] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1470.533199] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72cc3457-cff8-4877-a7f7-ba18ed0333d7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1470.538127] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1470.538127] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525fea00-89e2-fb08-1176-c57b57d235a3" [ 1470.538127] env[61970]: _type = "Task" [ 1470.538127] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1470.544986] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525fea00-89e2-fb08-1176-c57b57d235a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1470.928616] env[61970]: DEBUG oslo_concurrency.lockutils [req-e8cc27bb-d848-4f34-8d98-a453dac24d2f req-f48d6add-b976-4f23-a541-bbda735847a5 service nova] Releasing lock "refresh_cache-39117eed-03b2-4619-b8d6-04f09309a7ea" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1471.047972] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525fea00-89e2-fb08-1176-c57b57d235a3, 'name': SearchDatastore_Task, 'duration_secs': 0.007833} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1471.048746] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61f8435e-68c7-4a36-af5d-7e0e87c8af17 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1471.053840] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1471.053840] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5242b57d-0020-f3b9-28d2-41bec503773b" [ 1471.053840] env[61970]: _type = "Task" [ 1471.053840] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1471.060930] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5242b57d-0020-f3b9-28d2-41bec503773b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1471.564547] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]5242b57d-0020-f3b9-28d2-41bec503773b, 'name': SearchDatastore_Task, 'duration_secs': 0.008511} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1471.564847] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1471.565148] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 39117eed-03b2-4619-b8d6-04f09309a7ea/39117eed-03b2-4619-b8d6-04f09309a7ea.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1471.565410] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a0112de3-66ef-4051-af64-f36771670836 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1471.574222] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1471.574222] env[61970]: value = "task-1356382" [ 1471.574222] env[61970]: _type = "Task" [ 1471.574222] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1471.581373] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356382, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1472.084078] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356382, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497893} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1472.084442] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7ff5af44-3f73-4c84-82e5-a1543db26d03/7ff5af44-3f73-4c84-82e5-a1543db26d03.vmdk to [datastore2] 39117eed-03b2-4619-b8d6-04f09309a7ea/39117eed-03b2-4619-b8d6-04f09309a7ea.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1472.084516] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Extending root virtual disk to 1048576 {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1472.084770] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5f3ec366-5ba2-41bc-bd31-1d503f7fafe5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1472.090422] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1472.090422] env[61970]: value = "task-1356383" [ 1472.090422] env[61970]: _type = "Task" [ 1472.090422] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1472.097955] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356383, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1472.600106] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356383, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062745} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1472.600380] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Extended root virtual disk {{(pid=61970) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1472.601130] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4527e516-00e0-4b31-8a22-3f6c51b10200 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1472.622163] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 39117eed-03b2-4619-b8d6-04f09309a7ea/39117eed-03b2-4619-b8d6-04f09309a7ea.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1472.622399] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-129c1f33-3a15-4b7c-871e-2b63aa0ffc72 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1472.640522] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1472.640522] env[61970]: value = "task-1356384" [ 1472.640522] env[61970]: _type = "Task" [ 1472.640522] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1472.650535] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356384, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1473.150829] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356384, 'name': ReconfigVM_Task, 'duration_secs': 0.421799} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1473.151265] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 39117eed-03b2-4619-b8d6-04f09309a7ea/39117eed-03b2-4619-b8d6-04f09309a7ea.vmdk or device None with type sparse {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1473.151806] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4acaeb1a-5d87-44a2-a48b-caf9a5b7ba5f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1473.158322] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1473.158322] env[61970]: value = "task-1356385" [ 1473.158322] env[61970]: _type = "Task" [ 1473.158322] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1473.170959] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356385, 'name': Rename_Task} progress is 6%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1473.668509] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356385, 'name': Rename_Task, 'duration_secs': 0.144941} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1473.669213] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1473.669583] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1ebf5c4e-cfc8-467f-9ba0-07cef05cdbc2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1473.676632] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1473.676632] env[61970]: value = "task-1356386" [ 1473.676632] env[61970]: _type = "Task" [ 1473.676632] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1473.689439] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356386, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1474.187179] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356386, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1474.691037] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356386, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1475.189227] env[61970]: DEBUG oslo_vmware.api [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356386, 'name': PowerOnVM_Task, 'duration_secs': 1.034878} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1475.189606] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1475.189711] env[61970]: INFO nova.compute.manager [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Took 7.69 seconds to spawn the instance on the hypervisor. [ 1475.189891] env[61970]: DEBUG nova.compute.manager [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1475.190676] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67389f6-1490-4e17-9978-95846c53d2d8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1475.708907] env[61970]: INFO nova.compute.manager [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Took 12.36 seconds to build instance. [ 1476.211310] env[61970]: DEBUG oslo_concurrency.lockutils [None req-e3c59be3-08d1-445f-8035-7a080815698c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "39117eed-03b2-4619-b8d6-04f09309a7ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.867s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1476.603555] env[61970]: DEBUG oslo_concurrency.lockutils [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "39117eed-03b2-4619-b8d6-04f09309a7ea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1476.603874] env[61970]: DEBUG oslo_concurrency.lockutils [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "39117eed-03b2-4619-b8d6-04f09309a7ea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1476.604117] env[61970]: DEBUG oslo_concurrency.lockutils [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "39117eed-03b2-4619-b8d6-04f09309a7ea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1476.604320] env[61970]: DEBUG oslo_concurrency.lockutils [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "39117eed-03b2-4619-b8d6-04f09309a7ea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1476.604498] env[61970]: DEBUG oslo_concurrency.lockutils [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "39117eed-03b2-4619-b8d6-04f09309a7ea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1476.606730] env[61970]: INFO nova.compute.manager [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Terminating instance [ 1476.608456] env[61970]: DEBUG nova.compute.manager [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1476.608657] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1476.609504] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b99835-9705-4f28-9a5d-6bd9e033dcfb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1476.616929] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1476.617451] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5a39cae6-2493-4697-8bcf-953111ca9c5c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1476.623294] env[61970]: DEBUG oslo_vmware.api [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1476.623294] env[61970]: value = "task-1356387" [ 1476.623294] env[61970]: _type = "Task" [ 1476.623294] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1476.631068] env[61970]: DEBUG oslo_vmware.api [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356387, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1477.133459] env[61970]: DEBUG oslo_vmware.api [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356387, 'name': PowerOffVM_Task, 'duration_secs': 0.273851} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1477.133718] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1477.133911] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1477.134190] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-364caf5e-0cb3-4b53-8e83-7b8ce93b398a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1477.195376] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1477.195566] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1477.195749] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Deleting the datastore file [datastore2] 39117eed-03b2-4619-b8d6-04f09309a7ea {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1477.196017] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7399618c-0e66-479d-90bd-53d04ff86524 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1477.201917] env[61970]: DEBUG oslo_vmware.api [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1477.201917] env[61970]: value = "task-1356389" [ 1477.201917] env[61970]: _type = "Task" [ 1477.201917] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1477.209344] env[61970]: DEBUG oslo_vmware.api [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356389, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1477.711728] env[61970]: DEBUG oslo_vmware.api [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356389, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139103} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1477.712172] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1477.712219] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1477.712382] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1477.712563] env[61970]: INFO nova.compute.manager [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1477.712821] env[61970]: DEBUG oslo.service.loopingcall [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1477.713068] env[61970]: DEBUG nova.compute.manager [-] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1477.713174] env[61970]: DEBUG nova.network.neutron [-] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1477.958616] env[61970]: DEBUG nova.compute.manager [req-c54de11f-71d9-4042-9f89-7dc7628550b2 req-0aeb8d49-e35b-4413-84d2-4fce6fa11396 service nova] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Received event network-vif-deleted-27454272-523e-44c5-8c57-dd33bdd14c61 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1477.958800] env[61970]: INFO nova.compute.manager [req-c54de11f-71d9-4042-9f89-7dc7628550b2 req-0aeb8d49-e35b-4413-84d2-4fce6fa11396 service nova] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Neutron deleted interface 27454272-523e-44c5-8c57-dd33bdd14c61; detaching it from the instance and deleting it from the info cache [ 1477.958981] env[61970]: DEBUG nova.network.neutron [req-c54de11f-71d9-4042-9f89-7dc7628550b2 req-0aeb8d49-e35b-4413-84d2-4fce6fa11396 service nova] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1478.432038] env[61970]: DEBUG nova.network.neutron [-] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1478.461022] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-80b3b615-49cd-4941-9eb7-c8322acdc564 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1478.470814] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55118a7c-6e5d-4bb7-a196-b39a3ce95668 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1478.492840] env[61970]: DEBUG nova.compute.manager [req-c54de11f-71d9-4042-9f89-7dc7628550b2 req-0aeb8d49-e35b-4413-84d2-4fce6fa11396 service nova] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Detach interface failed, port_id=27454272-523e-44c5-8c57-dd33bdd14c61, reason: Instance 39117eed-03b2-4619-b8d6-04f09309a7ea could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1478.935043] env[61970]: INFO nova.compute.manager [-] [instance: 39117eed-03b2-4619-b8d6-04f09309a7ea] Took 1.22 seconds to deallocate network for instance. [ 1479.441873] env[61970]: DEBUG oslo_concurrency.lockutils [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1479.442151] env[61970]: DEBUG oslo_concurrency.lockutils [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1479.442393] env[61970]: DEBUG nova.objects.instance [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lazy-loading 'resources' on Instance uuid 39117eed-03b2-4619-b8d6-04f09309a7ea {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1479.986850] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d8ee81-f526-4699-8c3d-0cd1f3acf459 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1479.994441] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-830bf4b7-4e3b-4ece-a2db-f6b45d8dffa2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1480.023029] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-411b537f-78d4-4772-97d0-f371b819bd9c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1480.029407] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f4bf18-a6d5-4e6b-a8d5-6586737863df {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1480.043297] env[61970]: DEBUG nova.compute.provider_tree [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1480.546397] env[61970]: DEBUG nova.scheduler.client.report [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1481.051504] env[61970]: DEBUG oslo_concurrency.lockutils [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.609s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1481.073517] env[61970]: INFO nova.scheduler.client.report [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Deleted allocations for instance 39117eed-03b2-4619-b8d6-04f09309a7ea [ 1481.581801] env[61970]: DEBUG oslo_concurrency.lockutils [None req-053a470a-9151-46cd-b671-789300d8a3a7 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "39117eed-03b2-4619-b8d6-04f09309a7ea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.978s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1491.817279] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1491.817679] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1491.817775] env[61970]: INFO nova.compute.manager [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Shelving [ 1492.325067] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1492.325347] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5d00834c-6d78-4982-ac3f-8d5a01cd32f0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1492.333052] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1492.333052] env[61970]: value = "task-1356390" [ 1492.333052] env[61970]: _type = "Task" [ 1492.333052] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1492.341316] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356390, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1492.843137] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356390, 'name': PowerOffVM_Task, 'duration_secs': 0.182194} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1492.843403] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1492.844166] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820b3e56-d8b4-4670-a5f7-ed61ce4874e1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1492.864555] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba2002d0-ec22-4814-9176-e32eaf056166 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1493.375774] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Creating Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1493.376139] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-65bff714-8105-496b-b36a-08b8d095dd14 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1493.383738] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1493.383738] env[61970]: value = "task-1356391" [ 1493.383738] env[61970]: _type = "Task" [ 1493.383738] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1493.391484] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356391, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1493.893943] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356391, 'name': CreateSnapshot_Task, 'duration_secs': 0.459642} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1493.894383] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Created Snapshot of the VM instance {{(pid=61970) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1493.895036] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd6867f-8f10-4c36-924c-352d74dac7b1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1494.412989] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Creating linked-clone VM from snapshot {{(pid=61970) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1494.413312] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7b947d85-fcdd-4c2a-a9f3-b185c0eca564 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1494.421797] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1494.421797] env[61970]: value = "task-1356392" [ 1494.421797] env[61970]: _type = "Task" [ 1494.421797] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1494.429663] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356392, 'name': CloneVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1494.931749] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356392, 'name': CloneVM_Task} progress is 94%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1495.432474] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356392, 'name': CloneVM_Task, 'duration_secs': 0.943237} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1495.432743] env[61970]: INFO nova.virt.vmwareapi.vmops [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Created linked-clone VM from snapshot [ 1495.433493] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2007aa4-e803-48ce-87e4-d0c279947d44 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1495.440325] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Uploading image 6fcc5c9c-f2fc-44c6-b827-43f49842f732 {{(pid=61970) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1495.463121] env[61970]: DEBUG oslo_vmware.rw_handles [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1495.463121] env[61970]: value = "vm-288911" [ 1495.463121] env[61970]: _type = "VirtualMachine" [ 1495.463121] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1495.463374] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0627ee6f-ca61-492f-afd2-41586d5a17ad {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1495.470350] env[61970]: DEBUG oslo_vmware.rw_handles [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lease: (returnval){ [ 1495.470350] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52cb46dc-cd7d-e959-3d61-51a173fce56c" [ 1495.470350] env[61970]: _type = "HttpNfcLease" [ 1495.470350] env[61970]: } obtained for exporting VM: (result){ [ 1495.470350] env[61970]: value = "vm-288911" [ 1495.470350] env[61970]: _type = "VirtualMachine" [ 1495.470350] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1495.470590] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the lease: (returnval){ [ 1495.470590] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52cb46dc-cd7d-e959-3d61-51a173fce56c" [ 1495.470590] env[61970]: _type = "HttpNfcLease" [ 1495.470590] env[61970]: } to be ready. {{(pid=61970) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1495.476455] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1495.476455] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52cb46dc-cd7d-e959-3d61-51a173fce56c" [ 1495.476455] env[61970]: _type = "HttpNfcLease" [ 1495.476455] env[61970]: } is initializing. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1495.978682] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1495.978682] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52cb46dc-cd7d-e959-3d61-51a173fce56c" [ 1495.978682] env[61970]: _type = "HttpNfcLease" [ 1495.978682] env[61970]: } is ready. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1495.979195] env[61970]: DEBUG oslo_vmware.rw_handles [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1495.979195] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52cb46dc-cd7d-e959-3d61-51a173fce56c" [ 1495.979195] env[61970]: _type = "HttpNfcLease" [ 1495.979195] env[61970]: }. {{(pid=61970) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1495.979705] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56171db9-26eb-4e40-91ac-5e11999faa39 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1495.986842] env[61970]: DEBUG oslo_vmware.rw_handles [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528d8ac8-e252-158b-010d-44ab647ccc33/disk-0.vmdk from lease info. {{(pid=61970) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1495.987029] env[61970]: DEBUG oslo_vmware.rw_handles [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528d8ac8-e252-158b-010d-44ab647ccc33/disk-0.vmdk for reading. {{(pid=61970) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1496.075997] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-81f0ae45-55fc-4f6a-aa9f-e45922a730c7 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1503.213997] env[61970]: DEBUG oslo_vmware.rw_handles [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528d8ac8-e252-158b-010d-44ab647ccc33/disk-0.vmdk. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1503.215033] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f57d768f-6aa7-4fa8-a1da-1ac4a5cf386a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1503.221207] env[61970]: DEBUG oslo_vmware.rw_handles [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528d8ac8-e252-158b-010d-44ab647ccc33/disk-0.vmdk is in state: ready. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1503.221379] env[61970]: ERROR oslo_vmware.rw_handles [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528d8ac8-e252-158b-010d-44ab647ccc33/disk-0.vmdk due to incomplete transfer. [ 1503.221593] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-0e60f574-69e1-47d9-aa72-1886d40647b2 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1503.229220] env[61970]: DEBUG oslo_vmware.rw_handles [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528d8ac8-e252-158b-010d-44ab647ccc33/disk-0.vmdk. {{(pid=61970) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1503.229421] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Uploaded image 6fcc5c9c-f2fc-44c6-b827-43f49842f732 to the Glance image server {{(pid=61970) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1503.231616] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Destroying the VM {{(pid=61970) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1503.231845] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-bc486781-bae3-44d1-9d10-041fecf93da1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1503.237604] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1503.237604] env[61970]: value = "task-1356394" [ 1503.237604] env[61970]: _type = "Task" [ 1503.237604] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1503.244823] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356394, 'name': Destroy_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1503.748399] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356394, 'name': Destroy_Task, 'duration_secs': 0.374562} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1503.748658] env[61970]: INFO nova.virt.vmwareapi.vm_util [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Destroyed the VM [ 1503.748902] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Deleting Snapshot of the VM instance {{(pid=61970) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1503.749176] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-76273900-8c97-4aa6-9bf2-ee489e5f5140 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1503.757135] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1503.757135] env[61970]: value = "task-1356395" [ 1503.757135] env[61970]: _type = "Task" [ 1503.757135] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1503.764298] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356395, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1504.267282] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356395, 'name': RemoveSnapshot_Task, 'duration_secs': 0.339611} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1504.267594] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Deleted Snapshot of the VM instance {{(pid=61970) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1504.267881] env[61970]: DEBUG nova.compute.manager [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1504.268636] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f5a5eb1-3a6a-4a58-a430-dd18c25996ef {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1504.780295] env[61970]: INFO nova.compute.manager [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Shelve offloading [ 1504.781839] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1504.782101] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dae59454-245e-4913-9352-835cd7ab9f53 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1504.788953] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1504.788953] env[61970]: value = "task-1356396" [ 1504.788953] env[61970]: _type = "Task" [ 1504.788953] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1504.796617] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356396, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1505.260168] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1505.299024] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] VM already powered off {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1505.299415] env[61970]: DEBUG nova.compute.manager [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1505.299958] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817bb0ae-75d3-4a3a-a3ff-8339c0759a25 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1505.305512] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1505.305705] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquired lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1505.305883] env[61970]: DEBUG nova.network.neutron [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1506.016771] env[61970]: DEBUG nova.network.neutron [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Updating instance_info_cache with network_info: [{"id": "01289152-eff0-4751-94e5-732eaefc7cb0", "address": "fa:16:3e:d9:0e:c3", "network": {"id": "8c857731-1df4-4254-93c3-5a86f4a2dd2b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1262016675-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "92a836a268414a49a13fd816aa1c2407", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01289152-ef", "ovs_interfaceid": "01289152-eff0-4751-94e5-732eaefc7cb0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1506.519322] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Releasing lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1506.723021] env[61970]: DEBUG nova.compute.manager [req-5f4b232a-81ee-41cc-a185-a70303e49168 req-82ca7cdb-4d6b-40b6-aa19-be9d25e34064 service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Received event network-vif-unplugged-01289152-eff0-4751-94e5-732eaefc7cb0 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1506.723216] env[61970]: DEBUG oslo_concurrency.lockutils [req-5f4b232a-81ee-41cc-a185-a70303e49168 req-82ca7cdb-4d6b-40b6-aa19-be9d25e34064 service nova] Acquiring lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1506.723431] env[61970]: DEBUG oslo_concurrency.lockutils [req-5f4b232a-81ee-41cc-a185-a70303e49168 req-82ca7cdb-4d6b-40b6-aa19-be9d25e34064 service nova] Lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1506.723606] env[61970]: DEBUG oslo_concurrency.lockutils [req-5f4b232a-81ee-41cc-a185-a70303e49168 req-82ca7cdb-4d6b-40b6-aa19-be9d25e34064 service nova] Lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1506.723782] env[61970]: DEBUG nova.compute.manager [req-5f4b232a-81ee-41cc-a185-a70303e49168 req-82ca7cdb-4d6b-40b6-aa19-be9d25e34064 service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] No waiting events found dispatching network-vif-unplugged-01289152-eff0-4751-94e5-732eaefc7cb0 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1506.723952] env[61970]: WARNING nova.compute.manager [req-5f4b232a-81ee-41cc-a185-a70303e49168 req-82ca7cdb-4d6b-40b6-aa19-be9d25e34064 service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Received unexpected event network-vif-unplugged-01289152-eff0-4751-94e5-732eaefc7cb0 for instance with vm_state shelved and task_state shelving_offloading. [ 1506.750937] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1506.751816] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7acfb1f4-0ddf-49cd-8030-7c65ea49730f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1506.759599] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1506.759829] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9ad2a613-c6ab-4829-bb31-e8258d72ee22 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1506.829559] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1506.829798] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1506.829989] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Deleting the datastore file [datastore2] 99e6c9f1-6ecd-439b-a716-c378c18ec9bb {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1506.830283] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2881afb2-9356-48e1-8b30-b138a457c3b8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1506.836706] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1506.836706] env[61970]: value = "task-1356398" [ 1506.836706] env[61970]: _type = "Task" [ 1506.836706] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1506.844153] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356398, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1507.346759] env[61970]: DEBUG oslo_vmware.api [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356398, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12504} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1507.347019] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1507.347203] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1507.347380] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1507.371110] env[61970]: INFO nova.scheduler.client.report [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Deleted allocations for instance 99e6c9f1-6ecd-439b-a716-c378c18ec9bb [ 1507.875348] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1507.875755] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1507.875817] env[61970]: DEBUG nova.objects.instance [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lazy-loading 'resources' on Instance uuid 99e6c9f1-6ecd-439b-a716-c378c18ec9bb {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1508.378503] env[61970]: DEBUG nova.objects.instance [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lazy-loading 'numa_topology' on Instance uuid 99e6c9f1-6ecd-439b-a716-c378c18ec9bb {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1508.749140] env[61970]: DEBUG nova.compute.manager [req-35e843cc-c627-4025-a339-c8574ccfa7bb req-d4b7c149-25de-4961-a4d4-df003cf2fd5f service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Received event network-changed-01289152-eff0-4751-94e5-732eaefc7cb0 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1508.749351] env[61970]: DEBUG nova.compute.manager [req-35e843cc-c627-4025-a339-c8574ccfa7bb req-d4b7c149-25de-4961-a4d4-df003cf2fd5f service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Refreshing instance network info cache due to event network-changed-01289152-eff0-4751-94e5-732eaefc7cb0. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1508.749571] env[61970]: DEBUG oslo_concurrency.lockutils [req-35e843cc-c627-4025-a339-c8574ccfa7bb req-d4b7c149-25de-4961-a4d4-df003cf2fd5f service nova] Acquiring lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1508.749721] env[61970]: DEBUG oslo_concurrency.lockutils [req-35e843cc-c627-4025-a339-c8574ccfa7bb req-d4b7c149-25de-4961-a4d4-df003cf2fd5f service nova] Acquired lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1508.749888] env[61970]: DEBUG nova.network.neutron [req-35e843cc-c627-4025-a339-c8574ccfa7bb req-d4b7c149-25de-4961-a4d4-df003cf2fd5f service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Refreshing network info cache for port 01289152-eff0-4751-94e5-732eaefc7cb0 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1508.881171] env[61970]: DEBUG nova.objects.base [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Object Instance<99e6c9f1-6ecd-439b-a716-c378c18ec9bb> lazy-loaded attributes: resources,numa_topology {{(pid=61970) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1508.908068] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28fd5d7e-c0e0-486e-807a-7945f1484a6b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1508.915673] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f498770-dc96-422d-b0b1-d93dcbf4fb9f {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1508.943639] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd894652-02b1-49dc-86e2-d3b90782fc8c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1508.950218] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19ca393c-6756-41c3-bef4-f3f6b6074859 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1508.962520] env[61970]: DEBUG nova.compute.provider_tree [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1509.442658] env[61970]: DEBUG nova.network.neutron [req-35e843cc-c627-4025-a339-c8574ccfa7bb req-d4b7c149-25de-4961-a4d4-df003cf2fd5f service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Updated VIF entry in instance network info cache for port 01289152-eff0-4751-94e5-732eaefc7cb0. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1509.443085] env[61970]: DEBUG nova.network.neutron [req-35e843cc-c627-4025-a339-c8574ccfa7bb req-d4b7c149-25de-4961-a4d4-df003cf2fd5f service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Updating instance_info_cache with network_info: [{"id": "01289152-eff0-4751-94e5-732eaefc7cb0", "address": "fa:16:3e:d9:0e:c3", "network": {"id": "8c857731-1df4-4254-93c3-5a86f4a2dd2b", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1262016675-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "92a836a268414a49a13fd816aa1c2407", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap01289152-ef", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1509.465374] env[61970]: DEBUG nova.scheduler.client.report [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1509.599407] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1509.946015] env[61970]: DEBUG oslo_concurrency.lockutils [req-35e843cc-c627-4025-a339-c8574ccfa7bb req-d4b7c149-25de-4961-a4d4-df003cf2fd5f service nova] Releasing lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1509.970164] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.094s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1510.477958] env[61970]: DEBUG oslo_concurrency.lockutils [None req-9b268700-6978-473b-bb9f-95516dd41a8c tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 18.660s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1510.478832] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.880s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1510.479039] env[61970]: INFO nova.compute.manager [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Unshelving [ 1511.260021] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1511.260397] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1511.506898] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1511.507192] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1511.507408] env[61970]: DEBUG nova.objects.instance [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lazy-loading 'pci_requests' on Instance uuid 99e6c9f1-6ecd-439b-a716-c378c18ec9bb {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1511.763037] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1512.010938] env[61970]: DEBUG nova.objects.instance [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lazy-loading 'numa_topology' on Instance uuid 99e6c9f1-6ecd-439b-a716-c378c18ec9bb {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1512.514601] env[61970]: INFO nova.compute.claims [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1513.550474] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcc4e58a-0fed-423b-a60e-34e7f277da68 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1513.558155] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-152376b7-1d51-42bb-85e5-9402d7e72c4b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1513.588380] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f8fb8a-ead6-45bd-874d-7e226d16fd70 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1513.594943] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b29a0e-0b10-42aa-83b0-2a54338a2d4d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1513.607456] env[61970]: DEBUG nova.compute.provider_tree [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1514.111234] env[61970]: DEBUG nova.scheduler.client.report [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1514.616033] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.109s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1514.618201] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.855s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1514.618392] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1514.618552] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61970) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1514.619429] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c339c84c-390f-4341-921c-7bc79811a477 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1514.627455] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-940e7181-5c07-438e-ac37-425de3650b8a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1514.640624] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fb504fd-ca90-4c32-aed2-2fb75aca3e69 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1514.646552] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9929adc4-7b6c-4841-8441-7445760c2cbb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1514.649935] env[61970]: INFO nova.network.neutron [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Updating port 01289152-eff0-4751-94e5-732eaefc7cb0 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1514.677742] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181511MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61970) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1514.677887] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1514.678088] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1515.698930] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Instance 99e6c9f1-6ecd-439b-a716-c378c18ec9bb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61970) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1515.699221] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1515.699320] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61970) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1515.723609] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b818645-2b76-408e-9f94-2928830c5d5d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1515.730905] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59cd6314-fd9a-44eb-be5d-796e71a4208e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1515.759732] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e6e6558-48fb-41ed-bf0f-eee94bd3ed4a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1515.766567] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6906fa1-8040-4e48-86af-b77cabf9cfe4 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1515.779236] env[61970]: DEBUG nova.compute.provider_tree [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1515.998060] env[61970]: DEBUG nova.compute.manager [req-ed105275-58a0-45f3-a6ee-d03ca9a2a9c8 req-52d0e00c-0c88-4b99-a7ae-180f1e168976 service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Received event network-vif-plugged-01289152-eff0-4751-94e5-732eaefc7cb0 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1515.998626] env[61970]: DEBUG oslo_concurrency.lockutils [req-ed105275-58a0-45f3-a6ee-d03ca9a2a9c8 req-52d0e00c-0c88-4b99-a7ae-180f1e168976 service nova] Acquiring lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1515.998854] env[61970]: DEBUG oslo_concurrency.lockutils [req-ed105275-58a0-45f3-a6ee-d03ca9a2a9c8 req-52d0e00c-0c88-4b99-a7ae-180f1e168976 service nova] Lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1515.999046] env[61970]: DEBUG oslo_concurrency.lockutils [req-ed105275-58a0-45f3-a6ee-d03ca9a2a9c8 req-52d0e00c-0c88-4b99-a7ae-180f1e168976 service nova] Lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1515.999230] env[61970]: DEBUG nova.compute.manager [req-ed105275-58a0-45f3-a6ee-d03ca9a2a9c8 req-52d0e00c-0c88-4b99-a7ae-180f1e168976 service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] No waiting events found dispatching network-vif-plugged-01289152-eff0-4751-94e5-732eaefc7cb0 {{(pid=61970) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1515.999403] env[61970]: WARNING nova.compute.manager [req-ed105275-58a0-45f3-a6ee-d03ca9a2a9c8 req-52d0e00c-0c88-4b99-a7ae-180f1e168976 service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Received unexpected event network-vif-plugged-01289152-eff0-4751-94e5-732eaefc7cb0 for instance with vm_state shelved_offloaded and task_state spawning. [ 1516.076263] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1516.076440] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquired lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1516.076618] env[61970]: DEBUG nova.network.neutron [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1516.282557] env[61970]: DEBUG nova.scheduler.client.report [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1516.751426] env[61970]: DEBUG nova.network.neutron [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Updating instance_info_cache with network_info: [{"id": "01289152-eff0-4751-94e5-732eaefc7cb0", "address": "fa:16:3e:d9:0e:c3", "network": {"id": "8c857731-1df4-4254-93c3-5a86f4a2dd2b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1262016675-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "92a836a268414a49a13fd816aa1c2407", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01289152-ef", "ovs_interfaceid": "01289152-eff0-4751-94e5-732eaefc7cb0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1516.787482] env[61970]: DEBUG nova.compute.resource_tracker [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61970) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1516.787660] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.110s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1517.254315] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Releasing lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1517.283271] env[61970]: DEBUG nova.virt.hardware [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T13:36:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='452f1b50ce56e325a6613a3cca86027a',container_format='bare',created_at=2024-10-24T13:54:04Z,direct_url=,disk_format='vmdk',id=6fcc5c9c-f2fc-44c6-b827-43f49842f732,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-408116350-shelved',owner='92a836a268414a49a13fd816aa1c2407',properties=ImageMetaProps,protected=,size=31596032,status='active',tags=,updated_at=2024-10-24T13:54:16Z,virtual_size=,visibility=), allow threads: False {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1517.283521] env[61970]: DEBUG nova.virt.hardware [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Flavor limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1517.283683] env[61970]: DEBUG nova.virt.hardware [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Image limits 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1517.283866] env[61970]: DEBUG nova.virt.hardware [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Flavor pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1517.284027] env[61970]: DEBUG nova.virt.hardware [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Image pref 0:0:0 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1517.284187] env[61970]: DEBUG nova.virt.hardware [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61970) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1517.284396] env[61970]: DEBUG nova.virt.hardware [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1517.284559] env[61970]: DEBUG nova.virt.hardware [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1517.284730] env[61970]: DEBUG nova.virt.hardware [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Got 1 possible topologies {{(pid=61970) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1517.284895] env[61970]: DEBUG nova.virt.hardware [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1517.285085] env[61970]: DEBUG nova.virt.hardware [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61970) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1517.285940] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ecf99d0-bb1e-4f3a-8aa7-992638b98d09 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1517.294066] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f872c025-6111-4369-bfc7-9875bd1da25a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1517.307034] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d9:0e:c3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9c621a9c-66f5-426a-8aab-bd8b2e912106', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '01289152-eff0-4751-94e5-732eaefc7cb0', 'vif_model': 'vmxnet3'}] {{(pid=61970) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1517.314392] env[61970]: DEBUG oslo.service.loopingcall [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1517.314623] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Creating VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1517.314828] env[61970]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f5e5c5e-1158-41a0-bad8-0135e029c33b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1517.332210] env[61970]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1517.332210] env[61970]: value = "task-1356399" [ 1517.332210] env[61970]: _type = "Task" [ 1517.332210] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1517.339290] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356399, 'name': CreateVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1517.841417] env[61970]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356399, 'name': CreateVM_Task, 'duration_secs': 0.295809} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1517.841775] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Created VM on the ESX host {{(pid=61970) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1517.842193] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6fcc5c9c-f2fc-44c6-b827-43f49842f732" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1517.842368] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6fcc5c9c-f2fc-44c6-b827-43f49842f732" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1517.842744] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6fcc5c9c-f2fc-44c6-b827-43f49842f732" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1517.843017] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ddbf948-0292-4d6f-b7bc-5f32076b0fb0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1517.847015] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1517.847015] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525b1319-08c2-5532-b0ec-693968107f05" [ 1517.847015] env[61970]: _type = "Task" [ 1517.847015] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1517.854546] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]525b1319-08c2-5532-b0ec-693968107f05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1518.023425] env[61970]: DEBUG nova.compute.manager [req-16094c92-29f2-4005-8b5a-e0bf55de3a0d req-9c0a44b0-5bd2-4fa9-84fb-852b2061090c service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Received event network-changed-01289152-eff0-4751-94e5-732eaefc7cb0 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1518.023569] env[61970]: DEBUG nova.compute.manager [req-16094c92-29f2-4005-8b5a-e0bf55de3a0d req-9c0a44b0-5bd2-4fa9-84fb-852b2061090c service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Refreshing instance network info cache due to event network-changed-01289152-eff0-4751-94e5-732eaefc7cb0. {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1518.023788] env[61970]: DEBUG oslo_concurrency.lockutils [req-16094c92-29f2-4005-8b5a-e0bf55de3a0d req-9c0a44b0-5bd2-4fa9-84fb-852b2061090c service nova] Acquiring lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1518.023934] env[61970]: DEBUG oslo_concurrency.lockutils [req-16094c92-29f2-4005-8b5a-e0bf55de3a0d req-9c0a44b0-5bd2-4fa9-84fb-852b2061090c service nova] Acquired lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1518.024111] env[61970]: DEBUG nova.network.neutron [req-16094c92-29f2-4005-8b5a-e0bf55de3a0d req-9c0a44b0-5bd2-4fa9-84fb-852b2061090c service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Refreshing network info cache for port 01289152-eff0-4751-94e5-732eaefc7cb0 {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1518.357770] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6fcc5c9c-f2fc-44c6-b827-43f49842f732" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1518.358040] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Processing image 6fcc5c9c-f2fc-44c6-b827-43f49842f732 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1518.358287] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6fcc5c9c-f2fc-44c6-b827-43f49842f732/6fcc5c9c-f2fc-44c6-b827-43f49842f732.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1518.358441] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6fcc5c9c-f2fc-44c6-b827-43f49842f732/6fcc5c9c-f2fc-44c6-b827-43f49842f732.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1518.358618] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1518.358869] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-984ba5ca-0aca-4b3e-93d2-56b86fa3c05e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1518.368102] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1518.368327] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61970) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1518.369081] env[61970]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4356f8bd-c24d-4f43-8d99-bc7f2102ea2b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1518.376082] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1518.376082] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52723e86-d994-4529-32fa-61727f951ffd" [ 1518.376082] env[61970]: _type = "Task" [ 1518.376082] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1518.384955] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]52723e86-d994-4529-32fa-61727f951ffd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1518.702831] env[61970]: DEBUG nova.network.neutron [req-16094c92-29f2-4005-8b5a-e0bf55de3a0d req-9c0a44b0-5bd2-4fa9-84fb-852b2061090c service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Updated VIF entry in instance network info cache for port 01289152-eff0-4751-94e5-732eaefc7cb0. {{(pid=61970) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1518.703209] env[61970]: DEBUG nova.network.neutron [req-16094c92-29f2-4005-8b5a-e0bf55de3a0d req-9c0a44b0-5bd2-4fa9-84fb-852b2061090c service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Updating instance_info_cache with network_info: [{"id": "01289152-eff0-4751-94e5-732eaefc7cb0", "address": "fa:16:3e:d9:0e:c3", "network": {"id": "8c857731-1df4-4254-93c3-5a86f4a2dd2b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1262016675-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "92a836a268414a49a13fd816aa1c2407", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01289152-ef", "ovs_interfaceid": "01289152-eff0-4751-94e5-732eaefc7cb0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1518.782724] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1518.782930] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1518.783071] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Starting heal instance info cache {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1518.783192] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Rebuilding the list of instances to heal {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1518.886361] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Preparing fetch location {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1518.886756] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Fetch image to [datastore2] OSTACK_IMG_83ed09e8-4eda-4734-b782-d93be9402482/OSTACK_IMG_83ed09e8-4eda-4734-b782-d93be9402482.vmdk {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1518.886756] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Downloading stream optimized image 6fcc5c9c-f2fc-44c6-b827-43f49842f732 to [datastore2] OSTACK_IMG_83ed09e8-4eda-4734-b782-d93be9402482/OSTACK_IMG_83ed09e8-4eda-4734-b782-d93be9402482.vmdk on the data store datastore2 as vApp {{(pid=61970) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1518.886972] env[61970]: DEBUG nova.virt.vmwareapi.images [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Downloading image file data 6fcc5c9c-f2fc-44c6-b827-43f49842f732 to the ESX as VM named 'OSTACK_IMG_83ed09e8-4eda-4734-b782-d93be9402482' {{(pid=61970) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1518.951412] env[61970]: DEBUG oslo_vmware.rw_handles [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1518.951412] env[61970]: value = "resgroup-9" [ 1518.951412] env[61970]: _type = "ResourcePool" [ 1518.951412] env[61970]: }. {{(pid=61970) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1518.951691] env[61970]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-7c84a241-7c7a-494a-9372-6569bf6851ad {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1518.972556] env[61970]: DEBUG oslo_vmware.rw_handles [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lease: (returnval){ [ 1518.972556] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528e9ad1-6989-fb62-294d-fef392532b02" [ 1518.972556] env[61970]: _type = "HttpNfcLease" [ 1518.972556] env[61970]: } obtained for vApp import into resource pool (val){ [ 1518.972556] env[61970]: value = "resgroup-9" [ 1518.972556] env[61970]: _type = "ResourcePool" [ 1518.972556] env[61970]: }. {{(pid=61970) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1518.972895] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the lease: (returnval){ [ 1518.972895] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528e9ad1-6989-fb62-294d-fef392532b02" [ 1518.972895] env[61970]: _type = "HttpNfcLease" [ 1518.972895] env[61970]: } to be ready. {{(pid=61970) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1518.978608] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1518.978608] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528e9ad1-6989-fb62-294d-fef392532b02" [ 1518.978608] env[61970]: _type = "HttpNfcLease" [ 1518.978608] env[61970]: } is initializing. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1519.206593] env[61970]: DEBUG oslo_concurrency.lockutils [req-16094c92-29f2-4005-8b5a-e0bf55de3a0d req-9c0a44b0-5bd2-4fa9-84fb-852b2061090c service nova] Releasing lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1519.287052] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquiring lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1519.287352] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Acquired lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1519.287442] env[61970]: DEBUG nova.network.neutron [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Forcefully refreshing network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1519.287609] env[61970]: DEBUG nova.objects.instance [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Lazy-loading 'info_cache' on Instance uuid 99e6c9f1-6ecd-439b-a716-c378c18ec9bb {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1519.480925] env[61970]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1519.480925] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528e9ad1-6989-fb62-294d-fef392532b02" [ 1519.480925] env[61970]: _type = "HttpNfcLease" [ 1519.480925] env[61970]: } is ready. {{(pid=61970) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1519.481275] env[61970]: DEBUG oslo_vmware.rw_handles [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1519.481275] env[61970]: value = "session[5208e74c-dbbe-03c9-d237-66b24ffa2be9]528e9ad1-6989-fb62-294d-fef392532b02" [ 1519.481275] env[61970]: _type = "HttpNfcLease" [ 1519.481275] env[61970]: }. {{(pid=61970) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1519.481892] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa76fe9-34f1-4eae-9d18-51ed37c29e97 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1519.488833] env[61970]: DEBUG oslo_vmware.rw_handles [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52574fac-162b-7aa8-4a6a-c075ab368738/disk-0.vmdk from lease info. {{(pid=61970) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1519.489034] env[61970]: DEBUG oslo_vmware.rw_handles [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Creating HTTP connection to write to file with size = 31596032 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52574fac-162b-7aa8-4a6a-c075ab368738/disk-0.vmdk. {{(pid=61970) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1519.552062] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6166c0a2-c749-4a92-94dd-77db27fa5019 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1520.460430] env[61970]: DEBUG oslo_vmware.rw_handles [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Completed reading data from the image iterator. {{(pid=61970) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1520.460936] env[61970]: DEBUG oslo_vmware.rw_handles [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52574fac-162b-7aa8-4a6a-c075ab368738/disk-0.vmdk. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1520.461672] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce227fc-8856-4338-92b0-f94ad30ce75d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1520.469156] env[61970]: DEBUG oslo_vmware.rw_handles [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52574fac-162b-7aa8-4a6a-c075ab368738/disk-0.vmdk is in state: ready. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1520.469324] env[61970]: DEBUG oslo_vmware.rw_handles [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52574fac-162b-7aa8-4a6a-c075ab368738/disk-0.vmdk. {{(pid=61970) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1520.469571] env[61970]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-fdacd4f9-643c-4bdc-bb95-d4b050156969 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1520.705343] env[61970]: DEBUG oslo_vmware.rw_handles [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52574fac-162b-7aa8-4a6a-c075ab368738/disk-0.vmdk. {{(pid=61970) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1520.705580] env[61970]: INFO nova.virt.vmwareapi.images [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Downloaded image file data 6fcc5c9c-f2fc-44c6-b827-43f49842f732 [ 1520.706435] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8d0fe7-6c0c-4b1a-ae88-959c2a3132a8 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1520.721080] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-321de81d-6118-47e4-bc92-38dd866aa8b1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1520.750506] env[61970]: INFO nova.virt.vmwareapi.images [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] The imported VM was unregistered [ 1520.752745] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Caching image {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1520.752996] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Creating directory with path [datastore2] devstack-image-cache_base/6fcc5c9c-f2fc-44c6-b827-43f49842f732 {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1520.753287] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8a36d4b5-a16f-4e0f-95b4-72ce0cf70b7a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1520.762869] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Created directory with path [datastore2] devstack-image-cache_base/6fcc5c9c-f2fc-44c6-b827-43f49842f732 {{(pid=61970) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1520.763077] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_83ed09e8-4eda-4734-b782-d93be9402482/OSTACK_IMG_83ed09e8-4eda-4734-b782-d93be9402482.vmdk to [datastore2] devstack-image-cache_base/6fcc5c9c-f2fc-44c6-b827-43f49842f732/6fcc5c9c-f2fc-44c6-b827-43f49842f732.vmdk. {{(pid=61970) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1520.763319] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-f9748b12-fef3-4e37-97ed-ce499ed98702 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1520.769219] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1520.769219] env[61970]: value = "task-1356402" [ 1520.769219] env[61970]: _type = "Task" [ 1520.769219] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1520.776268] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356402, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1520.999777] env[61970]: DEBUG nova.network.neutron [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Updating instance_info_cache with network_info: [{"id": "01289152-eff0-4751-94e5-732eaefc7cb0", "address": "fa:16:3e:d9:0e:c3", "network": {"id": "8c857731-1df4-4254-93c3-5a86f4a2dd2b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1262016675-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "92a836a268414a49a13fd816aa1c2407", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01289152-ef", "ovs_interfaceid": "01289152-eff0-4751-94e5-732eaefc7cb0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1521.278999] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356402, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1521.503345] env[61970]: DEBUG oslo_concurrency.lockutils [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Releasing lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1521.503774] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Updated the network info_cache for instance {{(pid=61970) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1521.503774] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1521.503909] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1521.504527] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1521.504817] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1521.505104] env[61970]: DEBUG nova.compute.manager [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61970) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1521.779323] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356402, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1522.280524] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356402, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1522.781378] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356402, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1523.281538] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356402, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.180355} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1523.281796] env[61970]: INFO nova.virt.vmwareapi.ds_util [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_83ed09e8-4eda-4734-b782-d93be9402482/OSTACK_IMG_83ed09e8-4eda-4734-b782-d93be9402482.vmdk to [datastore2] devstack-image-cache_base/6fcc5c9c-f2fc-44c6-b827-43f49842f732/6fcc5c9c-f2fc-44c6-b827-43f49842f732.vmdk. [ 1523.281986] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Cleaning up location [datastore2] OSTACK_IMG_83ed09e8-4eda-4734-b782-d93be9402482 {{(pid=61970) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1523.282174] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_83ed09e8-4eda-4734-b782-d93be9402482 {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1523.282426] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8e837803-767c-4b49-a420-7cbd377b071b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1523.288281] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1523.288281] env[61970]: value = "task-1356403" [ 1523.288281] env[61970]: _type = "Task" [ 1523.288281] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1523.295374] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356403, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1523.798411] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356403, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.038337} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1523.798844] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1523.798844] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6fcc5c9c-f2fc-44c6-b827-43f49842f732/6fcc5c9c-f2fc-44c6-b827-43f49842f732.vmdk" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1523.799049] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6fcc5c9c-f2fc-44c6-b827-43f49842f732/6fcc5c9c-f2fc-44c6-b827-43f49842f732.vmdk to [datastore2] 99e6c9f1-6ecd-439b-a716-c378c18ec9bb/99e6c9f1-6ecd-439b-a716-c378c18ec9bb.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1523.799304] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9f253c80-d8ef-4718-8ac0-6980acbf9603 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1523.805915] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1523.805915] env[61970]: value = "task-1356404" [ 1523.805915] env[61970]: _type = "Task" [ 1523.805915] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1523.813021] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356404, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1524.316148] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356404, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1524.816721] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356404, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1525.317755] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356404, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1525.819238] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356404, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1526.319779] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356404, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.368624} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1526.320020] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6fcc5c9c-f2fc-44c6-b827-43f49842f732/6fcc5c9c-f2fc-44c6-b827-43f49842f732.vmdk to [datastore2] 99e6c9f1-6ecd-439b-a716-c378c18ec9bb/99e6c9f1-6ecd-439b-a716-c378c18ec9bb.vmdk {{(pid=61970) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1526.320791] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0feb3f27-a01d-42a6-9d47-30d80b72ebe9 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1526.341795] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 99e6c9f1-6ecd-439b-a716-c378c18ec9bb/99e6c9f1-6ecd-439b-a716-c378c18ec9bb.vmdk or device None with type streamOptimized {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1526.342056] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c19bac8-4add-46b7-b242-da221f743c8c {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1526.361043] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1526.361043] env[61970]: value = "task-1356405" [ 1526.361043] env[61970]: _type = "Task" [ 1526.361043] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1526.368097] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356405, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1526.872157] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356405, 'name': ReconfigVM_Task, 'duration_secs': 0.273714} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1526.872540] env[61970]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 99e6c9f1-6ecd-439b-a716-c378c18ec9bb/99e6c9f1-6ecd-439b-a716-c378c18ec9bb.vmdk or device None with type streamOptimized {{(pid=61970) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1526.873020] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-43c107de-6ad7-4b2b-93ac-23bff3b6a0bf {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1526.879291] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1526.879291] env[61970]: value = "task-1356406" [ 1526.879291] env[61970]: _type = "Task" [ 1526.879291] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1526.886101] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356406, 'name': Rename_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1527.389961] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356406, 'name': Rename_Task, 'duration_secs': 0.15411} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1527.390258] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Powering on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1527.390505] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d522285b-6d5c-4cdc-9451-219ddc5d379b {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1527.396754] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1527.396754] env[61970]: value = "task-1356407" [ 1527.396754] env[61970]: _type = "Task" [ 1527.396754] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1527.403863] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356407, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1527.906477] env[61970]: DEBUG oslo_vmware.api [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356407, 'name': PowerOnVM_Task, 'duration_secs': 0.420802} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1527.906858] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Powered on the VM {{(pid=61970) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1527.999507] env[61970]: DEBUG nova.compute.manager [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1528.000411] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c106ff19-e9ca-4048-996c-1afafae96b7a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1528.516575] env[61970]: DEBUG oslo_concurrency.lockutils [None req-3e2d15d0-443a-4014-ba53-d99b258952cd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.038s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1529.522944] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab93a045-2b27-4a95-9e70-32a308176aeb {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1529.530121] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-98c5a45b-cec2-4625-9283-a2bbf2ef5ccd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Suspending the VM {{(pid=61970) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1529.530371] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-9c1e1d58-42e0-4e3f-b52a-f537b2bc6fe0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1529.536682] env[61970]: DEBUG oslo_vmware.api [None req-98c5a45b-cec2-4625-9283-a2bbf2ef5ccd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1529.536682] env[61970]: value = "task-1356408" [ 1529.536682] env[61970]: _type = "Task" [ 1529.536682] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1529.545548] env[61970]: DEBUG oslo_vmware.api [None req-98c5a45b-cec2-4625-9283-a2bbf2ef5ccd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356408, 'name': SuspendVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1530.048906] env[61970]: DEBUG oslo_vmware.api [None req-98c5a45b-cec2-4625-9283-a2bbf2ef5ccd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356408, 'name': SuspendVM_Task} progress is 75%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1530.547542] env[61970]: DEBUG oslo_vmware.api [None req-98c5a45b-cec2-4625-9283-a2bbf2ef5ccd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356408, 'name': SuspendVM_Task, 'duration_secs': 0.547872} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1530.547935] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-98c5a45b-cec2-4625-9283-a2bbf2ef5ccd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Suspended the VM {{(pid=61970) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1530.548080] env[61970]: DEBUG nova.compute.manager [None req-98c5a45b-cec2-4625-9283-a2bbf2ef5ccd tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1530.548827] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-410a5b89-a421-43fe-ba34-eb973cb58e1d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.896606] env[61970]: INFO nova.compute.manager [None req-5da0c236-541c-42b2-93f9-a5fa381772ad tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Resuming [ 1531.897240] env[61970]: DEBUG nova.objects.instance [None req-5da0c236-541c-42b2-93f9-a5fa381772ad tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lazy-loading 'flavor' on Instance uuid 99e6c9f1-6ecd-439b-a716-c378c18ec9bb {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1532.904606] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5da0c236-541c-42b2-93f9-a5fa381772ad tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1532.905027] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5da0c236-541c-42b2-93f9-a5fa381772ad tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquired lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1532.905027] env[61970]: DEBUG nova.network.neutron [None req-5da0c236-541c-42b2-93f9-a5fa381772ad tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Building network info cache for instance {{(pid=61970) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1533.606505] env[61970]: DEBUG nova.network.neutron [None req-5da0c236-541c-42b2-93f9-a5fa381772ad tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Updating instance_info_cache with network_info: [{"id": "01289152-eff0-4751-94e5-732eaefc7cb0", "address": "fa:16:3e:d9:0e:c3", "network": {"id": "8c857731-1df4-4254-93c3-5a86f4a2dd2b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1262016675-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "92a836a268414a49a13fd816aa1c2407", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01289152-ef", "ovs_interfaceid": "01289152-eff0-4751-94e5-732eaefc7cb0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1534.109829] env[61970]: DEBUG oslo_concurrency.lockutils [None req-5da0c236-541c-42b2-93f9-a5fa381772ad tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Releasing lock "refresh_cache-99e6c9f1-6ecd-439b-a716-c378c18ec9bb" {{(pid=61970) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1534.110785] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26049135-38de-4178-b3f2-932ae0bdea10 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1534.117654] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5da0c236-541c-42b2-93f9-a5fa381772ad tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Resuming the VM {{(pid=61970) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1534.117896] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cacb4cfa-b2e5-4507-ab5a-3151e2136892 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1534.124611] env[61970]: DEBUG oslo_vmware.api [None req-5da0c236-541c-42b2-93f9-a5fa381772ad tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1534.124611] env[61970]: value = "task-1356409" [ 1534.124611] env[61970]: _type = "Task" [ 1534.124611] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1534.132515] env[61970]: DEBUG oslo_vmware.api [None req-5da0c236-541c-42b2-93f9-a5fa381772ad tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356409, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1534.639018] env[61970]: DEBUG oslo_vmware.api [None req-5da0c236-541c-42b2-93f9-a5fa381772ad tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356409, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1535.135752] env[61970]: DEBUG oslo_vmware.api [None req-5da0c236-541c-42b2-93f9-a5fa381772ad tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356409, 'name': PowerOnVM_Task, 'duration_secs': 0.54287} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1535.136677] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-5da0c236-541c-42b2-93f9-a5fa381772ad tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Resumed the VM {{(pid=61970) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1535.136780] env[61970]: DEBUG nova.compute.manager [None req-5da0c236-541c-42b2-93f9-a5fa381772ad tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Checking state {{(pid=61970) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1535.137475] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-966e42cc-495c-4afa-af1e-5ebdb2daba1e {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1540.155885] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1540.156266] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1540.156396] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1540.156576] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1540.156811] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1540.159090] env[61970]: INFO nova.compute.manager [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Terminating instance [ 1540.160830] env[61970]: DEBUG nova.compute.manager [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Start destroying the instance on the hypervisor. {{(pid=61970) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1540.161039] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Destroying instance {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1540.161863] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e03b6bb-fa36-4163-b724-12814266310a {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1540.169593] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Powering off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1540.169815] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b6a04e8-9929-48c2-a99a-d3de34d7fe34 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1540.177333] env[61970]: DEBUG oslo_vmware.api [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1540.177333] env[61970]: value = "task-1356410" [ 1540.177333] env[61970]: _type = "Task" [ 1540.177333] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1540.184863] env[61970]: DEBUG oslo_vmware.api [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356410, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1540.690680] env[61970]: DEBUG oslo_vmware.api [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356410, 'name': PowerOffVM_Task, 'duration_secs': 0.197444} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1540.691106] env[61970]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Powered off the VM {{(pid=61970) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1540.691370] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Unregistering the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1540.691700] env[61970]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1965ec81-7877-459a-9f13-823ede775ad1 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1540.768977] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Unregistered the VM {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1540.769245] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Deleting contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1540.769408] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Deleting the datastore file [datastore2] 99e6c9f1-6ecd-439b-a716-c378c18ec9bb {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1540.769689] env[61970]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-724c3b99-c387-42e6-b35a-53069e348cf0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1540.776573] env[61970]: DEBUG oslo_vmware.api [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for the task: (returnval){ [ 1540.776573] env[61970]: value = "task-1356412" [ 1540.776573] env[61970]: _type = "Task" [ 1540.776573] env[61970]: } to complete. {{(pid=61970) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1540.783856] env[61970]: DEBUG oslo_vmware.api [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356412, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1541.286469] env[61970]: DEBUG oslo_vmware.api [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Task: {'id': task-1356412, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131522} completed successfully. {{(pid=61970) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1541.286820] env[61970]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Deleted the datastore file {{(pid=61970) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1541.286945] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Deleted contents of the VM from datastore datastore2 {{(pid=61970) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1541.287085] env[61970]: DEBUG nova.virt.vmwareapi.vmops [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Instance destroyed {{(pid=61970) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1541.287267] env[61970]: INFO nova.compute.manager [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1541.287517] env[61970]: DEBUG oslo.service.loopingcall [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61970) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1541.287711] env[61970]: DEBUG nova.compute.manager [-] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Deallocating network for instance {{(pid=61970) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1541.287807] env[61970]: DEBUG nova.network.neutron [-] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] deallocate_for_instance() {{(pid=61970) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1541.508612] env[61970]: DEBUG nova.compute.manager [req-d52b9124-0413-4959-a210-ccc2032270e9 req-53e4402a-032d-4ad4-ba23-b9a5a4b97f3e service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Received event network-vif-deleted-01289152-eff0-4751-94e5-732eaefc7cb0 {{(pid=61970) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1541.508831] env[61970]: INFO nova.compute.manager [req-d52b9124-0413-4959-a210-ccc2032270e9 req-53e4402a-032d-4ad4-ba23-b9a5a4b97f3e service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Neutron deleted interface 01289152-eff0-4751-94e5-732eaefc7cb0; detaching it from the instance and deleting it from the info cache [ 1541.509011] env[61970]: DEBUG nova.network.neutron [req-d52b9124-0413-4959-a210-ccc2032270e9 req-53e4402a-032d-4ad4-ba23-b9a5a4b97f3e service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1541.986452] env[61970]: DEBUG nova.network.neutron [-] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Updating instance_info_cache with network_info: [] {{(pid=61970) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1542.012060] env[61970]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f182cc40-245c-4193-9fc0-220ca470230d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1542.021236] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47fea487-6eec-4a1e-be26-84db379936c0 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1542.042140] env[61970]: DEBUG nova.compute.manager [req-d52b9124-0413-4959-a210-ccc2032270e9 req-53e4402a-032d-4ad4-ba23-b9a5a4b97f3e service nova] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Detach interface failed, port_id=01289152-eff0-4751-94e5-732eaefc7cb0, reason: Instance 99e6c9f1-6ecd-439b-a716-c378c18ec9bb could not be found. {{(pid=61970) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1542.489404] env[61970]: INFO nova.compute.manager [-] [instance: 99e6c9f1-6ecd-439b-a716-c378c18ec9bb] Took 1.20 seconds to deallocate network for instance. [ 1542.995685] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1542.996008] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1542.996249] env[61970]: DEBUG nova.objects.instance [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lazy-loading 'resources' on Instance uuid 99e6c9f1-6ecd-439b-a716-c378c18ec9bb {{(pid=61970) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1543.534037] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a66a4eb6-2993-44fc-892b-c763c2f5383d {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1543.542518] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c61f5d92-a60e-4b37-8787-ad5ee6214a37 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1543.571521] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c309f595-9a1d-4517-8dd3-e877853604c5 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1543.578415] env[61970]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8e7e796-4221-439a-bf7e-d1a14914b622 {{(pid=61970) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1543.590779] env[61970]: DEBUG nova.compute.provider_tree [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Inventory has not changed in ProviderTree for provider: fc0f3e4e-5ef5-45e9-8335-cc68743405aa {{(pid=61970) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1544.093615] env[61970]: DEBUG nova.scheduler.client.report [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Inventory has not changed for provider fc0f3e4e-5ef5-45e9-8335-cc68743405aa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61970) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1544.598515] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.602s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1544.616920] env[61970]: INFO nova.scheduler.client.report [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Deleted allocations for instance 99e6c9f1-6ecd-439b-a716-c378c18ec9bb [ 1545.126201] env[61970]: DEBUG oslo_concurrency.lockutils [None req-2d626a60-8766-48d5-875a-2ce07a748e93 tempest-ServersNegativeTestJSON-1201342021 tempest-ServersNegativeTestJSON-1201342021-project-member] Lock "99e6c9f1-6ecd-439b-a716-c378c18ec9bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.970s {{(pid=61970) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1567.262648] env[61970]: DEBUG oslo_service.periodic_task [None req-7bcc8dd1-b623-4698-8e5d-cef1b21e0147 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61970) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}}